[ 470.501855] env[63175]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63175) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.502275] env[63175]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63175) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.503054] env[63175]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63175) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.503433] env[63175]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 470.595977] env[63175]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63175) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 470.606049] env[63175]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63175) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 471.217151] env[63175]: INFO nova.virt.driver [None req-5d20cc92-bfbb-4d68-aa50-88f54e7a951b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 471.287484] env[63175]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.287891] env[63175]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.288175] env[63175]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63175) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 474.387953] env[63175]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-59662ea1-fd6e-491c-8d58-638d630c598d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.404449] env[63175]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63175) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 474.404650] env[63175]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7d20d2d3-4211-45da-9031-5247a77f2538 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.428792] env[63175]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 691a0. [ 474.428956] env[63175]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.141s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.429433] env[63175]: INFO nova.virt.vmwareapi.driver [None req-5d20cc92-bfbb-4d68-aa50-88f54e7a951b None None] VMware vCenter version: 7.0.3 [ 474.432738] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c79c6e-85ed-4c0d-a9e2-922b07cc320d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.451134] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7e5975-7c5e-4dd5-a8f0-118306862530 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.454801] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ae6322-2646-49d7-bc05-4a670598b642 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.462695] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae055ccb-8ff5-4c03-aab0-8d0a497c70f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.475176] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5d10b7-9090-47fb-a61d-c5c1f91d0a92 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.480738] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28a6304-48da-4d60-b185-2d47f450f136 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.510107] env[63175]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-7248b949-b0bf-45db-8048-c405145adbec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.514708] env[63175]: DEBUG nova.virt.vmwareapi.driver [None req-5d20cc92-bfbb-4d68-aa50-88f54e7a951b None None] Extension org.openstack.compute already exists. {{(pid=63175) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 474.517312] env[63175]: INFO nova.compute.provider_config [None req-5d20cc92-bfbb-4d68-aa50-88f54e7a951b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 475.020621] env[63175]: DEBUG nova.context [None req-5d20cc92-bfbb-4d68-aa50-88f54e7a951b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),bcb15ecc-1652-4ca7-9bb4-d3efa0bc4706(cell1) {{(pid=63175) load_cells /opt/stack/nova/nova/context.py:464}} [ 475.022718] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.022982] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.023662] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.024109] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Acquiring lock "bcb15ecc-1652-4ca7-9bb4-d3efa0bc4706" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.024327] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Lock "bcb15ecc-1652-4ca7-9bb4-d3efa0bc4706" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.025349] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Lock "bcb15ecc-1652-4ca7-9bb4-d3efa0bc4706" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.046317] env[63175]: INFO dbcounter [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Registered counter for database nova_cell0 [ 475.054561] env[63175]: INFO dbcounter [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Registered counter for database nova_cell1 [ 475.057951] env[63175]: DEBUG oslo_db.sqlalchemy.engines [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63175) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.058333] env[63175]: DEBUG oslo_db.sqlalchemy.engines [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63175) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.063198] env[63175]: ERROR nova.db.main.api [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.063198] env[63175]: result = function(*args, **kwargs) [ 475.063198] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.063198] env[63175]: return func(*args, **kwargs) [ 475.063198] env[63175]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.063198] env[63175]: result = fn(*args, **kwargs) [ 475.063198] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.063198] env[63175]: return f(*args, **kwargs) [ 475.063198] env[63175]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.063198] env[63175]: return db.service_get_minimum_version(context, binaries) [ 475.063198] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.063198] env[63175]: _check_db_access() [ 475.063198] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.063198] env[63175]: stacktrace = ''.join(traceback.format_stack()) [ 475.063198] env[63175]: [ 475.064038] env[63175]: ERROR nova.db.main.api [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.064038] env[63175]: result = function(*args, **kwargs) [ 475.064038] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.064038] env[63175]: return func(*args, **kwargs) [ 475.064038] env[63175]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.064038] env[63175]: result = fn(*args, **kwargs) [ 475.064038] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.064038] env[63175]: return f(*args, **kwargs) [ 475.064038] env[63175]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.064038] env[63175]: return db.service_get_minimum_version(context, binaries) [ 475.064038] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.064038] env[63175]: _check_db_access() [ 475.064038] env[63175]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.064038] env[63175]: stacktrace = ''.join(traceback.format_stack()) [ 475.064038] env[63175]: [ 475.064461] env[63175]: WARNING nova.objects.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Failed to get minimum service version for cell bcb15ecc-1652-4ca7-9bb4-d3efa0bc4706 [ 475.064569] env[63175]: WARNING nova.objects.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 475.064994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Acquiring lock "singleton_lock" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 475.065184] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Acquired lock "singleton_lock" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 475.065409] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Releasing lock "singleton_lock" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 475.065735] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Full set of CONF: {{(pid=63175) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 475.065881] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ******************************************************************************** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 475.066030] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Configuration options gathered from: {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 475.066187] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 475.066384] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 475.066513] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ================================================================================ {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 475.066725] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] allow_resize_to_same_host = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.066896] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] arq_binding_timeout = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067036] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] backdoor_port = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067169] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] backdoor_socket = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067335] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] block_device_allocate_retries = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067497] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] block_device_allocate_retries_interval = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067670] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cert = self.pem {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.067830] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068013] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute_monitors = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068211] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] config_dir = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068386] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] config_drive_format = iso9660 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068517] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068681] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] config_source = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.068847] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] console_host = devstack {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069034] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] control_exchange = nova {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069219] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cpu_allocation_ratio = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069381] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] daemon = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069546] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] debug = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069698] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_access_ip_network_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.069861] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_availability_zone = nova {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070027] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_ephemeral_format = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070197] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_green_pool_size = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070439] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070609] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] default_schedule_zone = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070754] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] disk_allocation_ratio = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.070917] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] enable_new_services = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071104] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] enabled_apis = ['osapi_compute'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071277] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] enabled_ssl_apis = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071442] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] flat_injected = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071602] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] force_config_drive = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071764] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] force_raw_images = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.071932] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] graceful_shutdown_timeout = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.072107] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] heal_instance_info_cache_interval = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.072331] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] host = cpu-1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.072510] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.072670] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.072828] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073050] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073234] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_build_timeout = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073417] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_delete_interval = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073589] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_format = [instance: %(uuid)s] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073767] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_name_template = instance-%08x {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.073917] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_usage_audit = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074097] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_usage_audit_period = month {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074266] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074429] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074589] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] internal_service_availability_zone = internal {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074743] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] key = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.074901] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] live_migration_retry_count = 30 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075074] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_color = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075287] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_config_append = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075397] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075551] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_dir = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075705] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075828] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_options = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.075987] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_rotate_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076194] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_rotate_interval_type = days {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076368] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] log_rotation_type = none {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076506] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076634] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076800] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.076962] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077102] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077269] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] long_rpc_timeout = 1800 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077428] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_concurrent_builds = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077582] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_concurrent_live_migrations = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077737] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_concurrent_snapshots = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.077896] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_local_block_devices = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078087] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_logfile_count = 30 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078266] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] max_logfile_size_mb = 200 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078427] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] maximum_instance_delete_attempts = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078592] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metadata_listen = 0.0.0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078756] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metadata_listen_port = 8775 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.078924] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metadata_workers = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.079120] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] migrate_max_retries = -1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.079301] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] mkisofs_cmd = genisoimage {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.079509] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.079641] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] my_ip = 10.180.1.21 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.079843] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080017] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] network_allocate_retries = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080206] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080374] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080535] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] osapi_compute_listen_port = 8774 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080699] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] osapi_compute_unique_server_name_scope = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.080870] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] osapi_compute_workers = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081039] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] password_length = 12 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081207] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] periodic_enable = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081365] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] periodic_fuzzy_delay = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081530] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] pointer_model = usbtablet {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081693] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] preallocate_images = none {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081849] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] publish_errors = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.081978] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] pybasedir = /opt/stack/nova {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082145] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ram_allocation_ratio = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082305] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rate_limit_burst = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082470] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rate_limit_except_level = CRITICAL {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082627] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rate_limit_interval = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082783] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reboot_timeout = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.082937] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reclaim_instance_interval = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083122] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] record = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083315] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reimage_timeout_per_gb = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083483] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] report_interval = 120 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083644] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rescue_timeout = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083800] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reserved_host_cpus = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.083956] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reserved_host_disk_mb = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084126] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reserved_host_memory_mb = 512 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084289] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] reserved_huge_pages = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084446] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] resize_confirm_window = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084604] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] resize_fs_using_block_device = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084762] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] resume_guests_state_on_host_boot = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.084928] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085103] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] rpc_response_timeout = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085266] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] run_external_periodic_tasks = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085431] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] running_deleted_instance_action = reap {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085587] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085745] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] running_deleted_instance_timeout = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.085904] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler_instance_sync_interval = 120 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086099] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_down_time = 720 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086284] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] servicegroup_driver = db {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086440] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] shell_completion = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086600] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] shelved_offload_time = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086758] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] shelved_poll_interval = 3600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.086920] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] shutdown_timeout = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087090] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] source_is_ipv6 = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087253] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ssl_only = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087499] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087668] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] sync_power_state_interval = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087827] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] sync_power_state_pool_size = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.087995] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] syslog_log_facility = LOG_USER {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088189] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] tempdir = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088356] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] timeout_nbd = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088521] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] transport_url = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088678] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] update_resources_interval = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088835] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_cow_images = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.088992] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_eventlog = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089162] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_journal = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089322] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_json = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089476] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_rootwrap_daemon = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089630] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_stderr = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089782] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] use_syslog = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.089931] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vcpu_pin_set = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090106] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plugging_is_fatal = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090274] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plugging_timeout = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090438] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] virt_mkfs = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090598] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] volume_usage_poll_interval = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090758] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] watch_log_file = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.090923] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] web = /usr/share/spice-html5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.091123] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.091303] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.091468] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.091639] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_concurrency.disable_process_locking = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.091926] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092122] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092293] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092464] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092631] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092795] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.092974] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.auth_strategy = keystone {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.093157] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.compute_link_prefix = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.093333] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.093507] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.dhcp_domain = novalocal {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.093676] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.enable_instance_password = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.093838] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.glance_link_prefix = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094009] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094187] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094347] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.instance_list_per_project_cells = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094504] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.list_records_by_skipping_down_cells = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094663] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.local_metadata_per_cell = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094829] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.max_limit = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.094995] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.metadata_cache_expiration = 15 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.095180] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.neutron_default_tenant_id = default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.095356] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.response_validation = warn {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.095523] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.use_neutron_default_nets = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.095687] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.095846] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096026] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096218] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096390] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_dynamic_targets = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096550] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_jsonfile_path = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096724] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.096914] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.backend = dogpile.cache.memcached {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097092] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.backend_argument = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097271] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.config_prefix = cache.oslo {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097437] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.dead_timeout = 60.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097598] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.debug_cache_backend = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097759] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.enable_retry_client = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.097918] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.enable_socket_keepalive = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098124] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.enabled = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098306] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.enforce_fips_mode = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098503] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.expiration_time = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098632] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.hashclient_retry_attempts = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098798] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.098957] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_dead_retry = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099126] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_password = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099293] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099452] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099611] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_pool_maxsize = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099771] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.099928] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_sasl_enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100117] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100290] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100475] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.memcache_username = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100645] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.proxies = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100809] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_db = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.100965] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_password = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101149] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101329] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101498] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_server = localhost:6379 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101661] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_socket_timeout = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101819] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.redis_username = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.101982] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.retry_attempts = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.102163] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.retry_delay = 0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.102366] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.socket_keepalive_count = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.102545] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.socket_keepalive_idle = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.102711] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.socket_keepalive_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.102874] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.tls_allowed_ciphers = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103043] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.tls_cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103209] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.tls_certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103372] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.tls_enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103527] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cache.tls_keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103695] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.103867] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.auth_type = password {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104037] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104217] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104423] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104623] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104792] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.cross_az_attach = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.104958] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.debug = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105140] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.endpoint_template = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105314] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.http_retries = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105478] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105639] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105808] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.os_region_name = RegionOne {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.105972] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106145] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cinder.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106320] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106476] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.cpu_dedicated_set = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106632] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.cpu_shared_set = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106795] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.image_type_exclude_list = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.106953] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107128] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107293] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107453] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107619] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107779] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.resource_provider_association_refresh = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.107940] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.108182] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.shutdown_retry_interval = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.108322] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.108501] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] conductor.workers = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.108681] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] console.allowed_origins = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.108841] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] console.ssl_ciphers = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109030] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] console.ssl_minimum_version = default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109191] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] consoleauth.enforce_session_timeout = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109363] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] consoleauth.token_ttl = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109533] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109689] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.109854] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110018] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110177] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110334] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110492] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110647] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110803] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.110959] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111127] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111285] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111440] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111606] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.service_type = accelerator {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111765] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.111920] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112089] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112253] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112432] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112592] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] cyborg.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112767] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.backend = sqlalchemy {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.112932] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.connection = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113108] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.connection_debug = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113279] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.connection_parameters = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113445] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.connection_recycle_time = 3600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113604] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.connection_trace = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113762] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.db_inc_retry_interval = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.113922] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.db_max_retries = 20 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114093] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.db_max_retry_interval = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114260] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.db_retry_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114417] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.max_overflow = 50 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114574] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.max_pool_size = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114730] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.max_retries = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.114893] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115060] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.mysql_wsrep_sync_wait = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115223] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.pool_timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115380] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.retry_interval = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115537] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.slave_connection = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115692] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.sqlite_synchronous = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.115848] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] database.use_db_reconnect = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.116075] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.backend = sqlalchemy {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.116351] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.connection = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.116553] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.connection_debug = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.116731] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.connection_parameters = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.116899] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.connection_recycle_time = 3600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117076] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.connection_trace = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117246] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.db_inc_retry_interval = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117411] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.db_max_retries = 20 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117574] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.db_max_retry_interval = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117737] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.db_retry_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.117897] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.max_overflow = 50 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118085] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.max_pool_size = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118289] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.max_retries = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118426] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118586] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118744] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.pool_timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.118907] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.retry_interval = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119088] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.slave_connection = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119251] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] api_database.sqlite_synchronous = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119428] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] devices.enabled_mdev_types = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119605] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119775] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.119937] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ephemeral_storage_encryption.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120115] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120293] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.api_servers = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120456] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120615] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120782] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.120941] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121114] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121280] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.debug = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121444] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.default_trusted_certificate_ids = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121606] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.enable_certificate_validation = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121769] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.enable_rbd_download = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.121933] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122110] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122273] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122428] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122582] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122742] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.num_retries = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.122907] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.rbd_ceph_conf = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123077] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.rbd_connect_timeout = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123248] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.rbd_pool = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123412] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.rbd_user = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123570] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123725] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.123880] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.124159] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.service_type = image {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.124353] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.124515] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.124674] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.124833] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125021] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125193] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.verify_glance_signatures = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125356] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] glance.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125523] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] guestfs.debug = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125692] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.125854] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126017] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126175] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126339] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126497] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126653] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126807] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.126968] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127139] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127299] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127456] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127612] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127769] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.127923] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128121] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.service_type = shared-file-system {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128297] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.share_apply_policy_timeout = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128461] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128618] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128771] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.128929] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.129125] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.129289] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] manila.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.129459] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] mks.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.129815] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130015] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.manager_interval = 2400 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130198] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.precache_concurrency = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130368] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.remove_unused_base_images = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130536] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130702] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.130875] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] image_cache.subdirectory_name = _base {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131062] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.api_max_retries = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131234] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.api_retry_interval = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131391] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131550] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131706] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.131861] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132029] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132198] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.conductor_group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132359] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132516] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132672] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132832] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.132985] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133160] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133314] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133476] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.peer_list = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133631] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133785] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.133946] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.serial_console_state_timeout = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134113] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134283] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.service_type = baremetal {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134439] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.shard = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134598] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134750] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.134904] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135070] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135254] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135416] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ironic.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135592] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135761] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] key_manager.fixed_key = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.135939] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136133] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.barbican_api_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136305] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.barbican_endpoint = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136478] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.barbican_endpoint_type = public {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136636] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.barbican_region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136815] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.136970] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137147] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137308] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137464] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137626] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.number_of_retries = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137785] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.retry_delay = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.137944] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.send_service_user_token = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138153] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138361] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138511] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.verify_ssl = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138643] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican.verify_ssl_path = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138810] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.138969] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.139179] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.139382] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.139551] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.139715] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.139873] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140047] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140218] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] barbican_service_user.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140381] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.approle_role_id = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140542] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.approle_secret_id = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140712] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.kv_mountpoint = secret {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.140870] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.kv_path = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141046] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.kv_version = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141211] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.namespace = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141371] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.root_token_id = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141534] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.ssl_ca_crt_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141705] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.timeout = 60.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.141865] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.use_ssl = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142044] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142221] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142382] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142540] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142696] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.142857] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143024] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143189] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143347] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143510] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143673] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.143861] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144043] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144189] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144349] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144506] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144685] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.service_type = identity {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.144855] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145027] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145196] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145354] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145534] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145693] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] keystone.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.145881] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.ceph_mount_options = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.146634] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.146834] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.connection_uri = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147013] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_mode = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147197] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147368] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_models = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147539] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_power_governor_high = performance {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147708] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.147874] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_power_management = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148080] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148272] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.device_detach_attempts = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148440] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.device_detach_timeout = 20 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148654] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.disk_cachemodes = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148774] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.disk_prefix = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.148943] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.enabled_perf_events = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.149144] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.file_backed_memory = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.149415] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.gid_maps = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.149591] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.hw_disk_discard = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.149756] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.hw_machine_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.149930] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_rbd_ceph_conf = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150113] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150286] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150456] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_rbd_glance_store_name = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150625] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_rbd_pool = rbd {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150794] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_type = default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.150954] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.images_volume_group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151132] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.inject_key = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151301] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.inject_partition = -2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151463] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.inject_password = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151626] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.iscsi_iface = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151787] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.iser_use_multipath = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.151949] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.152126] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.152291] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_downtime = 500 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.152475] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.152706] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.152884] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_inbound_addr = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153064] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153236] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153396] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_scheme = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153571] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_timeout_action = abort {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153731] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_tunnelled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.153886] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_uri = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.154060] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.live_migration_with_native_tls = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.154228] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.max_queues = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.154395] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.154642] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.154807] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.nfs_mount_options = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155115] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155297] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155463] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155624] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155789] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.155960] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_pcie_ports = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.156179] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.156359] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.pmem_namespaces = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.156521] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.quobyte_client_cfg = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.156818] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.156992] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157174] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157342] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157503] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rbd_secret_uuid = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157660] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rbd_user = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157824] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.157996] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.158190] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rescue_image_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.158355] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rescue_kernel_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.158515] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rescue_ramdisk_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.158684] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.158846] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.rx_queue_size = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.159028] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.smbfs_mount_options = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.159305] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.159478] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.snapshot_compression = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.159640] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.snapshot_image_format = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.159857] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160040] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.sparse_logical_volumes = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160211] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.swtpm_enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160380] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.swtpm_group = tss {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160545] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.swtpm_user = tss {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160711] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.sysinfo_serial = unique {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.160870] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.tb_cache_size = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161038] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.tx_queue_size = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161208] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.uid_maps = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161368] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.use_virtio_for_bridges = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161535] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.virt_type = kvm {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161702] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.volume_clear = zero {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.161859] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.volume_clear_size = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162034] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.volume_use_multipath = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162200] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_cache_path = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162368] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162534] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162699] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.162864] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.163158] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.163340] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.vzstorage_mount_user = stack {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.163507] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.163682] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.163854] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.auth_type = password {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164021] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164187] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164350] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164508] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164665] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164832] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.default_floating_pool = public {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.164989] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165163] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.extension_sync_interval = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165325] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.http_retries = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165482] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165637] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165790] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.165957] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166212] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166321] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.ovs_bridge = br-int {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166485] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.physnets = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166651] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.region_name = RegionOne {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166809] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.166975] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.service_metadata_proxy = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167145] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167314] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.service_type = network {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167479] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167636] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167790] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.167947] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.168169] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.168338] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] neutron.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.168513] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] notifications.bdms_in_notifications = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.168686] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] notifications.default_level = INFO {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.168856] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] notifications.notification_format = unversioned {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169032] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] notifications.notify_on_state_change = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169216] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169393] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] pci.alias = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169564] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] pci.device_spec = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169729] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] pci.report_in_placement = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.169903] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170087] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.auth_type = password {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170261] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170429] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170577] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170741] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.170901] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171075] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171240] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.default_domain_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171397] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.default_domain_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171554] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.domain_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171710] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.domain_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.171865] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172036] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172202] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172357] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172512] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172675] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.password = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172833] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.project_domain_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.172999] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.project_domain_name = Default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.173183] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.project_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.173356] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.project_name = service {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.173524] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.region_name = RegionOne {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.173686] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.173846] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174017] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.service_type = placement {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174185] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174345] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174503] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174661] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.system_scope = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174818] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.174976] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.trust_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175147] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.user_domain_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175319] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.user_domain_name = Default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175477] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.user_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175648] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.username = nova {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175827] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.175990] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] placement.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.176221] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.cores = 20 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.176402] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.count_usage_from_placement = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.176574] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.176750] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.injected_file_content_bytes = 10240 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.176921] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.injected_file_path_length = 255 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177099] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.injected_files = 5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177273] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.instances = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177441] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.key_pairs = 100 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177607] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.metadata_items = 128 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177773] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.ram = 51200 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.177938] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.recheck_quota = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178144] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.server_group_members = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178324] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] quota.server_groups = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178498] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178666] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178830] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.image_metadata_prefilter = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.178994] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.179177] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.max_attempts = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.179343] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.max_placement_results = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.179509] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.179673] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.179835] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180018] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] scheduler.workers = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180201] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180372] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180551] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180724] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.180889] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181065] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181236] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181423] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181589] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.host_subset_size = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181751] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.181907] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182076] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182245] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.isolated_hosts = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182407] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.isolated_images = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182566] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182723] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.182883] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183053] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.pci_in_placement = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183222] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183395] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183557] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183716] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.183874] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184042] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184245] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.track_instance_changes = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184457] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184635] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metrics.required = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184800] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metrics.weight_multiplier = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.184961] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.185140] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] metrics.weight_setting = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.185458] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.185629] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.185801] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.port_range = 10000:20000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.185969] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.186180] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.186353] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] serial_console.serialproxy_port = 6083 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.186520] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.186689] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.auth_type = password {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.186847] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187008] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187178] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187387] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187574] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187748] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.send_service_user_token = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.187912] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.188103] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] service_user.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.188288] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.agent_enabled = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.188453] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.188755] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.188947] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189132] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.html5proxy_port = 6082 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189297] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.image_compression = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189453] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.jpeg_compression = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189608] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.playback_compression = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189766] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.require_secure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.189932] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.server_listen = 127.0.0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190119] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190282] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.streaming_mode = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190439] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] spice.zlib_compression = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190603] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] upgrade_levels.baseapi = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190771] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] upgrade_levels.compute = auto {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.190928] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] upgrade_levels.conductor = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191093] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] upgrade_levels.scheduler = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191262] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191422] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191576] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191730] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.191888] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192058] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192219] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192377] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192532] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vendordata_dynamic_auth.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192705] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.api_retry_count = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.192865] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.ca_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193044] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193218] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.cluster_name = testcl1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193382] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.connection_pool_size = 10 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193538] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.console_delay_seconds = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193703] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.datastore_regex = ^datastore.* {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.193906] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194086] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.host_password = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194254] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.host_port = 443 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194419] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.host_username = administrator@vsphere.local {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194583] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.insecure = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194739] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.integration_bridge = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.194900] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.maximum_objects = 100 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195069] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.pbm_default_policy = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195233] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.pbm_enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195386] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.pbm_wsdl_location = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195551] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195706] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.serial_port_proxy_uri = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.195860] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.serial_port_service_uri = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196054] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.task_poll_interval = 0.5 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196257] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.use_linked_clone = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196432] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.vnc_keymap = en-us {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196598] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.vnc_port = 5900 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196761] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vmware.vnc_port_total = 10000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.196947] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.auth_schemes = ['none'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.197136] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.197435] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.197618] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.197790] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.novncproxy_port = 6080 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.197964] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.server_listen = 127.0.0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.198223] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.198422] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.vencrypt_ca_certs = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.198586] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.vencrypt_client_cert = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.198746] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vnc.vencrypt_client_key = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.198928] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199107] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_deep_image_inspection = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199281] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199441] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199602] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199764] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.disable_rootwrap = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.199923] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.enable_numa_live_migration = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200095] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200261] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200420] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200616] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.libvirt_disable_apic = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200789] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.200952] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201129] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201293] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201454] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201612] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201770] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.201926] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202095] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202266] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202447] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202616] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.client_socket_timeout = 900 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202780] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.default_pool_size = 1000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.202946] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.keep_alive = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203125] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.max_header_line = 16384 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203290] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203449] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.ssl_ca_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203606] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.ssl_cert_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203764] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.ssl_key_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.203926] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.tcp_keepidle = 600 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.204115] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.204289] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] zvm.ca_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.204492] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] zvm.cloud_connector_url = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.204782] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.204959] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] zvm.reachable_timeout = 300 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.205158] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.enforce_new_defaults = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.205546] env[63175]: WARNING oslo_config.cfg [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 475.205729] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.enforce_scope = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.205906] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.policy_default_rule = default {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206119] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206315] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.policy_file = policy.yaml {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206487] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206650] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206809] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.206966] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207148] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207320] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207493] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207665] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.connection_string = messaging:// {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207831] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.enabled = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.207997] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.es_doc_type = notification {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.208203] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.es_scroll_size = 10000 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.208380] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.es_scroll_time = 2m {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.208544] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.filter_error_trace = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.208712] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.hmac_keys = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.208878] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.sentinel_service_name = mymaster {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209055] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.socket_timeout = 0.1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209224] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.trace_requests = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209385] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler.trace_sqlalchemy = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209561] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler_jaeger.process_tags = {} {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209718] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler_jaeger.service_name_prefix = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.209879] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] profiler_otlp.service_name_prefix = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210051] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] remote_debug.host = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210215] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] remote_debug.port = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210392] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210550] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210710] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.210869] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211037] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211203] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211360] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211517] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211674] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211841] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.211997] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.212177] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.212346] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.212512] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.212680] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.212844] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213011] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213192] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213354] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213514] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213678] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.213842] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214014] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214192] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214354] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214515] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214676] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.214837] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215019] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215187] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215357] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215524] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215685] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.215852] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.216048] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.216253] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.216531] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.216731] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_notifications.retry = -1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.216921] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217113] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217291] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.auth_section = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217456] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.auth_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217611] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.cafile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217768] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.certfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.217928] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.collect_timing = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218123] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.connect_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218298] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.connect_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218455] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.endpoint_id = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218613] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.endpoint_override = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218774] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.insecure = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.218929] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.keyfile = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219134] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.max_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219346] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.min_version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219512] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.region_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219677] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.retriable_status_codes = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219834] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.service_name = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.219991] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.service_type = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220173] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.split_loggers = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220337] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.status_code_retries = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220494] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.status_code_retry_delay = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220650] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.timeout = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220807] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.valid_interfaces = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.220960] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_limit.version = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221138] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_reports.file_event_handler = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221305] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221463] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] oslo_reports.log_dir = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221634] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221791] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.221947] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222121] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222286] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222442] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222610] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222766] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.222923] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223097] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223266] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223421] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] vif_plug_ovs_privileged.user = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223592] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223770] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.223940] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.224160] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.224346] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.224513] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.224679] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.224839] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225028] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225207] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.isolate_vif = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225377] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225543] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225713] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.225881] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226068] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] os_vif_ovs.per_port_bridge = False {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226253] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.capabilities = [21] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226414] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226569] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.helper_command = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226733] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.226896] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227060] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] privsep_osbrick.user = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227240] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227397] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.group = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227552] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.helper_command = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227715] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.227873] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.228128] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] nova_sys_admin.user = None {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.228202] env[63175]: DEBUG oslo_service.service [None req-efe5e6c5-b6c5-4717-ae50-56a46502e1d0 None None] ******************************************************************************** {{(pid=63175) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 475.228688] env[63175]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 475.732539] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Getting list of instances from cluster (obj){ [ 475.732539] env[63175]: value = "domain-c8" [ 475.732539] env[63175]: _type = "ClusterComputeResource" [ 475.732539] env[63175]: } {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 475.733692] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2702749-8360-4b64-bce6-12f10c76bd14 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.742554] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Got total of 0 instances {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 475.743089] env[63175]: WARNING nova.virt.vmwareapi.driver [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 475.743563] env[63175]: INFO nova.virt.node [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Generated node identity f10bd0b1-85ed-4db3-a7f2-6e16e785219e [ 475.743792] env[63175]: INFO nova.virt.node [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Wrote node identity f10bd0b1-85ed-4db3-a7f2-6e16e785219e to /opt/stack/data/n-cpu-1/compute_id [ 476.247065] env[63175]: WARNING nova.compute.manager [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Compute nodes ['f10bd0b1-85ed-4db3-a7f2-6e16e785219e'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 477.253057] env[63175]: INFO nova.compute.manager [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 478.261116] env[63175]: WARNING nova.compute.manager [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 478.261116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.261116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.261116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.261116] env[63175]: DEBUG nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 478.261116] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cccff3-3e23-48f4-b0c0-80992ad19bc1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.268677] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f8a113-2973-433a-be7f-dbfb1659c74e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.283522] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1859945-fe1e-4cbe-9602-d69a7656a463 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.289904] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34342c9a-d2e1-40ad-821b-950afdabdce3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.318471] env[63175]: DEBUG nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181549MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 478.318809] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.319218] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.823123] env[63175]: WARNING nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] No compute node record for cpu-1:f10bd0b1-85ed-4db3-a7f2-6e16e785219e: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host f10bd0b1-85ed-4db3-a7f2-6e16e785219e could not be found. [ 479.326065] env[63175]: INFO nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: f10bd0b1-85ed-4db3-a7f2-6e16e785219e [ 480.835759] env[63175]: DEBUG nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 480.835759] env[63175]: DEBUG nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 480.983799] env[63175]: INFO nova.scheduler.client.report [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] [req-1a270209-ce0f-4d0e-8f1e-111b9b0bebe8] Created resource provider record via placement API for resource provider with UUID f10bd0b1-85ed-4db3-a7f2-6e16e785219e and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 481.001023] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d0f2ab-af8c-4238-ace4-388399a1d562 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.007301] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3549a3-936f-408b-8170-b2899841baa8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.036802] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3a23c0-7122-411a-91c7-acceb7dd0ae1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.043603] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4cdb31-a04c-410e-acd3-4d806854da64 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.056281] env[63175]: DEBUG nova.compute.provider_tree [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.604021] env[63175]: DEBUG nova.scheduler.client.report [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Updated inventory for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 481.604021] env[63175]: DEBUG nova.compute.provider_tree [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Updating resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e generation from 0 to 1 during operation: update_inventory {{(pid=63175) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 481.604021] env[63175]: DEBUG nova.compute.provider_tree [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.652669] env[63175]: DEBUG nova.compute.provider_tree [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Updating resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e generation from 1 to 2 during operation: update_traits {{(pid=63175) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 482.160033] env[63175]: DEBUG nova.compute.resource_tracker [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 482.160033] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.839s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 482.160033] env[63175]: DEBUG nova.service [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Creating RPC server for service compute {{(pid=63175) start /opt/stack/nova/nova/service.py:186}} [ 482.171867] env[63175]: DEBUG nova.service [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] Join ServiceGroup membership for this service compute {{(pid=63175) start /opt/stack/nova/nova/service.py:203}} [ 482.172340] env[63175]: DEBUG nova.servicegroup.drivers.db [None req-1f8e01b2-66b3-4f4e-833d-d4ed951af11a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63175) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 502.174121] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_power_states {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 502.677094] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Getting list of instances from cluster (obj){ [ 502.677094] env[63175]: value = "domain-c8" [ 502.677094] env[63175]: _type = "ClusterComputeResource" [ 502.677094] env[63175]: } {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 502.678275] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef96171-7f98-4af4-9c96-6a65a9d724ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.687205] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Got total of 0 instances {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 502.687430] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 502.687721] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Getting list of instances from cluster (obj){ [ 502.687721] env[63175]: value = "domain-c8" [ 502.687721] env[63175]: _type = "ClusterComputeResource" [ 502.687721] env[63175]: } {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 502.688581] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c847b5-ae37-4c4d-9e74-5b72ab1a11fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.698793] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Got total of 0 instances {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 518.389290] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquiring lock "0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.389646] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.898203] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 519.416331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "24ee74d8-33d2-4efa-97de-a642c89a3461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.416591] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.452851] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.453145] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.455732] env[63175]: INFO nova.compute.claims [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.880570] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquiring lock "a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.880817] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.920296] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 520.384969] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 520.450242] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.533675] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquiring lock "18acf9cd-0661-48c0-8fd2-b80edba4dfa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.534403] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "18acf9cd-0661-48c0-8fd2-b80edba4dfa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.557073] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c8d20b-91c5-4b91-8d2f-3449e853ee59 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.567492] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1e14db-5160-469a-b4d6-17a0a156cd69 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.599179] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36a5846-072f-46a4-aa20-907be6aeb66e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.607117] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89264785-04bd-4e35-9008-4a9937b0b77e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.624880] env[63175]: DEBUG nova.compute.provider_tree [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.914322] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.039924] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 521.050508] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "56b3ba41-0406-4b69-89af-b00e87feaa89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.050975] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.128059] env[63175]: DEBUG nova.scheduler.client.report [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 521.197623] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "380a6278-22d4-485b-b9cb-ccb1ac07d4a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.197623] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "380a6278-22d4-485b-b9cb-ccb1ac07d4a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.556920] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 521.571657] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.633962] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.636738] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 521.640361] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.188s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.640361] env[63175]: INFO nova.compute.claims [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.699951] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 521.886026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquiring lock "1561f155-ff79-4790-a164-43f08230e229" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.886026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "1561f155-ff79-4790-a164-43f08230e229" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.099817] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.148696] env[63175]: DEBUG nova.compute.utils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.150280] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 522.150421] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 522.228885] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.390252] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 522.618310] env[63175]: DEBUG nova.policy [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd972d7c8bf0343fc94a625e6608d7f76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49f9410f671844d7a166593c851ebfa2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.661626] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 522.809790] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6d9a2a-a0b0-4b86-a026-07299d1a9fcd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.819382] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac6f71b-3b5f-449e-a60f-2b501b8042da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.856760] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e669dd1-4f2b-4b12-aaf3-547018811c32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.865785] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de23045-a9c3-4348-a039-4736791b1d42 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.882212] env[63175]: DEBUG nova.compute.provider_tree [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.915641] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.385365] env[63175]: DEBUG nova.scheduler.client.report [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 523.679169] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 523.726729] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.728222] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.728481] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.728758] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.728938] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.729189] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.729526] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.729599] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.730024] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.730212] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.730459] env[63175]: DEBUG nova.virt.hardware [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.738682] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dc6afd-782b-4098-b2d5-3d9de6689bc5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.750200] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a96394-86b2-4bca-977d-4fee91d1d0be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.772546] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5e839b-f920-488a-a04f-089eddf07f5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.884928] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Successfully created port: e4fa27c5-f657-463e-b046-8fc23264bd9d {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.890868] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.891160] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 523.896199] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.979s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.896199] env[63175]: INFO nova.compute.claims [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.401035] env[63175]: DEBUG nova.compute.utils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 524.404257] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 524.449701] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquiring lock "0e0f0535-d539-4572-8752-e3e6a4b79b8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.449933] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "0e0f0535-d539-4572-8752-e3e6a4b79b8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.906722] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 524.952140] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 525.115039] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95003082-3f44-470a-a1cb-709c822d1471 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.126687] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268f184d-4ba9-4cc6-bef2-868f343b8c94 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.169311] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd01405-9abc-489c-8e72-30377e368fb6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.176854] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e0ec88-0332-4172-aebc-35ca7a38c4ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.191558] env[63175]: DEBUG nova.compute.provider_tree [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.495204] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.695485] env[63175]: DEBUG nova.scheduler.client.report [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 525.925983] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 525.976131] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.977379] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.978159] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.978159] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.978159] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.978159] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.978332] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.978514] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.978746] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.978947] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.979159] env[63175]: DEBUG nova.virt.hardware [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.980394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a3d830-32eb-415a-9a11-f592070d9ef5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.997018] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da077226-12e9-447e-95d7-0a59238a153b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.011048] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 526.022066] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 526.022431] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5225e5cb-4986-4af1-9823-a4547f03b57e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.033381] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created folder: OpenStack in parent group-v4. [ 526.033569] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating folder: Project (18cf2e2f47374cb69f7612d1991960f5). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 526.033882] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e28e4b93-1c3d-4db7-bb70-dbcca7c72afc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.048046] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created folder: Project (18cf2e2f47374cb69f7612d1991960f5) in parent group-v268956. [ 526.048046] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating folder: Instances. Parent ref: group-v268957. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 526.048046] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9d3eae7-b23e-4cc0-81df-693e35483026 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.055391] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created folder: Instances in parent group-v268957. [ 526.055845] env[63175]: DEBUG oslo.service.loopingcall [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.056199] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 526.056511] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-757a37f0-d2a3-43d6-9db3-248cfb63137d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.074949] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 526.074949] env[63175]: value = "task-1247834" [ 526.074949] env[63175]: _type = "Task" [ 526.074949] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.088124] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247834, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.202804] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.203291] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 526.207544] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.636s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.209051] env[63175]: INFO nova.compute.claims [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.593105] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247834, 'name': CreateVM_Task, 'duration_secs': 0.305238} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.593105] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 526.593105] env[63175]: DEBUG oslo_vmware.service [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20aff90-7e2d-4cf8-a4f4-57867b3970d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.599913] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.600103] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.600763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 526.601076] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a77502a-5869-459f-b9e0-b5078eb3c1ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.605644] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 526.605644] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5206ddaf-f08c-dda2-9aa6-837f95424aec" [ 526.605644] env[63175]: _type = "Task" [ 526.605644] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.615743] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5206ddaf-f08c-dda2-9aa6-837f95424aec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.716563] env[63175]: DEBUG nova.compute.utils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.722628] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 526.722714] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 526.844178] env[63175]: DEBUG nova.policy [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee3902f94caa4dbabf8c9e343f913c09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '636a00e4279f4aeca559455158a701fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.051969] env[63175]: ERROR nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 527.051969] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.051969] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.051969] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.051969] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.051969] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.051969] env[63175]: ERROR nova.compute.manager raise self.value [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.051969] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.051969] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.051969] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.054523] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.054523] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.054523] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 527.054523] env[63175]: ERROR nova.compute.manager [ 527.054523] env[63175]: Traceback (most recent call last): [ 527.054523] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.054523] env[63175]: listener.cb(fileno) [ 527.054523] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.054523] env[63175]: result = function(*args, **kwargs) [ 527.054523] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.054523] env[63175]: return func(*args, **kwargs) [ 527.054523] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 527.054523] env[63175]: raise e [ 527.054523] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.054523] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 527.054523] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.054523] env[63175]: created_port_ids = self._update_ports_for_instance( [ 527.054523] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.054523] env[63175]: with excutils.save_and_reraise_exception(): [ 527.054523] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.054523] env[63175]: self.force_reraise() [ 527.054523] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.054523] env[63175]: raise self.value [ 527.054523] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.054523] env[63175]: updated_port = self._update_port( [ 527.054523] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.054523] env[63175]: _ensure_no_port_binding_failure(port) [ 527.054523] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.054523] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.055341] env[63175]: nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 527.055341] env[63175]: Removing descriptor: 15 [ 527.055341] env[63175]: ERROR nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Traceback (most recent call last): [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] yield resources [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.driver.spawn(context, instance, image_meta, [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.055341] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] vm_ref = self.build_virtual_machine(instance, [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] for vif in network_info: [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self._sync_wrapper(fn, *args, **kwargs) [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.wait() [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self[:] = self._gt.wait() [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self._exit_event.wait() [ 527.055623] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] result = hub.switch() [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self.greenlet.switch() [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] result = function(*args, **kwargs) [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return func(*args, **kwargs) [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise e [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] nwinfo = self.network_api.allocate_for_instance( [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.055929] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] created_port_ids = self._update_ports_for_instance( [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] with excutils.save_and_reraise_exception(): [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.force_reraise() [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise self.value [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] updated_port = self._update_port( [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] _ensure_no_port_binding_failure(port) [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.056253] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise exception.PortBindingFailed(port_id=port['id']) [ 527.056533] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 527.056533] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] [ 527.056533] env[63175]: INFO nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Terminating instance [ 527.121026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.121245] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 527.121705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.122353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.122433] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 527.122893] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f0743ba-82de-4880-935b-0defdb77f3d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.142065] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 527.142962] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 527.143081] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee4fa57-3298-407e-bb62-6e3d7b7b67bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.150782] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ee0ae01-ba2d-41da-8c17-c9d7e5e69818 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.158159] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 527.158159] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526c0b09-f300-f602-3938-d5d8a21d26be" [ 527.158159] env[63175]: _type = "Task" [ 527.158159] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.167988] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526c0b09-f300-f602-3938-d5d8a21d26be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.223378] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 527.397816] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62cda73-8753-407b-b8f5-94d1080ff28a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.408102] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27684baa-231a-4ec7-9fb4-15caf1dd2683 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.440314] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c2bc5f-6fee-4c05-aab5-d7f90539189d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.450867] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebace1a-97b9-4110-b872-ade8ad32b268 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.471734] env[63175]: DEBUG nova.compute.provider_tree [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.549075] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Successfully created port: edfa6f97-302b-41e5-8c84-a3add567630a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.560338] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquiring lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.561997] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquired lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.561997] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 527.671757] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Preparing fetch location {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 527.671757] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating directory with path [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 527.671757] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e337a25-7bc7-42d0-bd74-8aa8e1b3560d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.691771] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created directory with path [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 527.691771] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Fetch image to [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 527.691771] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Downloading image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk on the data store datastore2 {{(pid=63175) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 527.691905] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e570a658-a3b2-418d-8c96-c62623fc041f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.706560] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319f56c4-4e66-4cc2-a588-d05d91fe2b12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.716949] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1dda05-a99a-4129-b87e-5434a14c7f9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.755416] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555405c5-6229-4e23-aa83-0b19197e4534 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.762335] env[63175]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-054493e9-252a-4922-9814-6377f7d78ce0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.858826] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Downloading image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to the data store datastore2 {{(pid=63175) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 527.917926] env[63175]: DEBUG oslo_vmware.rw_handles [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63175) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 527.984880] env[63175]: DEBUG nova.scheduler.client.report [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 528.110728] env[63175]: DEBUG nova.compute.manager [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Received event network-changed-e4fa27c5-f657-463e-b046-8fc23264bd9d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 528.110728] env[63175]: DEBUG nova.compute.manager [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Refreshing instance network info cache due to event network-changed-e4fa27c5-f657-463e-b046-8fc23264bd9d. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 528.110728] env[63175]: DEBUG oslo_concurrency.lockutils [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] Acquiring lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.136905] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.261423] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 528.288105] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.290018] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.290018] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.290018] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.290018] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.290018] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.290225] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.290225] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.290225] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.290225] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.290225] env[63175]: DEBUG nova.virt.hardware [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.291116] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e715a53e-ed57-4a87-ad27-f7ce92633edd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.306113] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12814e1-8f05-4cbd-8530-7cd575877944 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.316214] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.493692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.494204] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 528.497675] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.398s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.499401] env[63175]: INFO nova.compute.claims [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.565801] env[63175]: DEBUG oslo_vmware.rw_handles [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Completed reading data from the image iterator. {{(pid=63175) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 528.565801] env[63175]: DEBUG oslo_vmware.rw_handles [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 528.705177] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquiring lock "9595c38c-3699-4b89-bf37-b2e435886878" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.705412] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "9595c38c-3699-4b89-bf37-b2e435886878" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.713141] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Downloaded image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk on the data store datastore2 {{(pid=63175) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 528.714872] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Caching image {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 528.715133] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copying Virtual Disk [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk to [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.715708] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7e5e544-3a48-4078-9cd7-507a77cf7987 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.727091] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 528.727091] env[63175]: value = "task-1247835" [ 528.727091] env[63175]: _type = "Task" [ 528.727091] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.740305] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.820439] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Releasing lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.820934] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 528.821192] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 528.821696] env[63175]: DEBUG oslo_concurrency.lockutils [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] Acquired lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.821888] env[63175]: DEBUG nova.network.neutron [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Refreshing network info cache for port e4fa27c5-f657-463e-b046-8fc23264bd9d {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 528.824102] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3a35fae-82d0-4ebb-878a-de47ffd4708c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.834473] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c49ddf-e36b-4842-97c6-c3e435da8321 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.864603] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52 could not be found. [ 528.865299] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 528.865299] env[63175]: INFO nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Took 0.04 seconds to destroy the instance on the hypervisor. [ 528.865606] env[63175]: DEBUG oslo.service.loopingcall [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.866272] env[63175]: DEBUG nova.compute.manager [-] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 528.866272] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 528.914850] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.008771] env[63175]: DEBUG nova.compute.utils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.014094] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 529.014291] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.207590] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 529.239472] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247835, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.274074] env[63175]: DEBUG nova.policy [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fffc81a470a543a7aeef1f6bae74e09d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248562f05fef4335a24ff4d93f2dbbc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.388996] env[63175]: DEBUG nova.network.neutron [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.420301] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.514738] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 529.568999] env[63175]: DEBUG nova.network.neutron [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.683597] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b647c2-0380-41a2-b2f3-a860d8f0bc39 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.693159] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60852e0d-c52f-4804-a326-94d76e1b0b12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.727766] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd21872-2103-4544-b4a6-a62d769f9769 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.742268] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529d23eb-50f4-40d0-a3d1-e71c8d7714d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.760515] env[63175]: DEBUG nova.compute.provider_tree [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.763964] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.764408] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685009} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.764870] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copied Virtual Disk [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk to [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 529.764983] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleting the datastore file [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.765187] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8df46d03-7b93-4527-8cb7-8e4def22249f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.771596] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 529.771596] env[63175]: value = "task-1247836" [ 529.771596] env[63175]: _type = "Task" [ 529.771596] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.779813] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.924880] env[63175]: INFO nova.compute.manager [-] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Took 1.06 seconds to deallocate network for instance. [ 529.930124] env[63175]: DEBUG nova.compute.claims [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.930331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.073590] env[63175]: DEBUG oslo_concurrency.lockutils [req-f57fa7bb-af61-4be5-a527-04f1dedc1946 req-47f16ae2-861e-4633-8daa-1e86d484c708 service nova] Releasing lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.268595] env[63175]: DEBUG nova.scheduler.client.report [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 530.281804] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022635} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.286254] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 530.286520] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Moving file from [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508. {{(pid=63175) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 530.286836] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-f167393f-2f90-485c-a3f8-e88d87a61135 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.294705] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 530.294705] env[63175]: value = "task-1247837" [ 530.294705] env[63175]: _type = "Task" [ 530.294705] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.307201] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247837, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.527941] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 530.566107] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.566506] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.567031] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.567031] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.567180] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.567421] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.567590] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.567788] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.568022] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.568268] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.568518] env[63175]: DEBUG nova.virt.hardware [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.569968] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb18d4b7-0248-4985-a054-0cb0d8b4961d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.589802] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f923410c-8829-43e0-8a67-56b3cd4d2d04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.618463] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquiring lock "596f899b-18f2-42f0-968e-6e9159f1d4d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.618742] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "596f899b-18f2-42f0-968e-6e9159f1d4d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.621746] env[63175]: ERROR nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 530.621746] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 530.621746] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.621746] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.621746] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.621746] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.621746] env[63175]: ERROR nova.compute.manager raise self.value [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.621746] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 530.621746] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.621746] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 530.622920] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.622920] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 530.622920] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 530.622920] env[63175]: ERROR nova.compute.manager [ 530.622920] env[63175]: Traceback (most recent call last): [ 530.622920] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 530.622920] env[63175]: listener.cb(fileno) [ 530.622920] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.622920] env[63175]: result = function(*args, **kwargs) [ 530.622920] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.622920] env[63175]: return func(*args, **kwargs) [ 530.622920] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 530.622920] env[63175]: raise e [ 530.622920] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 530.622920] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 530.622920] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.622920] env[63175]: created_port_ids = self._update_ports_for_instance( [ 530.622920] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.622920] env[63175]: with excutils.save_and_reraise_exception(): [ 530.622920] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.622920] env[63175]: self.force_reraise() [ 530.622920] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.622920] env[63175]: raise self.value [ 530.622920] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.622920] env[63175]: updated_port = self._update_port( [ 530.622920] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.622920] env[63175]: _ensure_no_port_binding_failure(port) [ 530.622920] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.622920] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 530.623777] env[63175]: nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 530.623777] env[63175]: Removing descriptor: 17 [ 530.623777] env[63175]: ERROR nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Traceback (most recent call last): [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] yield resources [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.driver.spawn(context, instance, image_meta, [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.623777] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] vm_ref = self.build_virtual_machine(instance, [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] for vif in network_info: [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self._sync_wrapper(fn, *args, **kwargs) [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.wait() [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self[:] = self._gt.wait() [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self._exit_event.wait() [ 530.624229] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] result = hub.switch() [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self.greenlet.switch() [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] result = function(*args, **kwargs) [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return func(*args, **kwargs) [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise e [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] nwinfo = self.network_api.allocate_for_instance( [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 530.624562] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] created_port_ids = self._update_ports_for_instance( [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] with excutils.save_and_reraise_exception(): [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.force_reraise() [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise self.value [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] updated_port = self._update_port( [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] _ensure_no_port_binding_failure(port) [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.624919] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise exception.PortBindingFailed(port_id=port['id']) [ 530.625303] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 530.625303] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] [ 530.625303] env[63175]: INFO nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Terminating instance [ 530.661099] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.661486] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.661627] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 530.661700] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 530.775689] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.775689] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 530.776312] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.548s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.778304] env[63175]: INFO nova.compute.claims [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.808307] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247837, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.02796} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.808586] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] File moved {{(pid=63175) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 530.810518] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Cleaning up location [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 530.810518] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleting the datastore file [datastore2] vmware_temp/09d47fd2-3ff3-4e38-bf15-ef1908711f76 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 530.810518] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d950b8d7-fe83-4df9-b629-c34a1a9bfe7b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.820473] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 530.820473] env[63175]: value = "task-1247838" [ 530.820473] env[63175]: _type = "Task" [ 530.820473] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.829308] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.883986] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Successfully created port: 0d58a188-6ddc-4364-9486-421c86294015 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.122582] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 531.127156] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquiring lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.127421] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquired lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.127608] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 531.169681] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 531.169988] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 531.170183] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 531.170322] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 531.170446] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 531.170562] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 531.170803] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.171232] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.172833] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.173136] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.173357] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.173550] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.173849] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 531.173903] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.282740] env[63175]: DEBUG nova.compute.utils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.287688] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 531.332388] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022374} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.332631] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 531.335263] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40bc255e-3c21-4e06-8a9d-e059e0129f9e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.341129] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 531.341129] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c456ba-3b08-65e5-4719-39c351b4c7e6" [ 531.341129] env[63175]: _type = "Task" [ 531.341129] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.352761] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c456ba-3b08-65e5-4719-39c351b4c7e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.361951] env[63175]: DEBUG nova.compute.manager [req-b4e3ad0b-7658-4dce-8a16-8cd487708f4b req-94397f89-36d0-4cb1-9a75-b72e1633adb9 service nova] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Received event network-vif-deleted-e4fa27c5-f657-463e-b046-8fc23264bd9d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 531.677430] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.681021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.789018] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 531.853546] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c456ba-3b08-65e5-4719-39c351b4c7e6, 'name': SearchDatastore_Task, 'duration_secs': 0.011588} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.859473] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.859998] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 531.861827] env[63175]: DEBUG nova.compute.manager [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Received event network-changed-edfa6f97-302b-41e5-8c84-a3add567630a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 531.861827] env[63175]: DEBUG nova.compute.manager [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Refreshing instance network info cache due to event network-changed-edfa6f97-302b-41e5-8c84-a3add567630a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 531.861827] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] Acquiring lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.864766] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb91f7be-9412-4a39-9d7a-1609e91de764 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.871479] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 531.871479] env[63175]: value = "task-1247839" [ 531.871479] env[63175]: _type = "Task" [ 531.871479] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.886831] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.948078] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.074394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16d7cb9-e7aa-460f-b562-8a0e808f092b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.086738] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef8fbd6-a177-437a-886d-81263c9d67ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.125906] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59193ad-e963-4536-af21-65ebb36f4448 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.135934] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bf3039-62bd-4a43-96d7-31e55c4d1993 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.152559] env[63175]: DEBUG nova.compute.provider_tree [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.175164] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.383925] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247839, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.656143] env[63175]: DEBUG nova.scheduler.client.report [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 532.677817] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Releasing lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.678351] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 532.678541] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.679529] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] Acquired lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.679722] env[63175]: DEBUG nova.network.neutron [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Refreshing network info cache for port edfa6f97-302b-41e5-8c84-a3add567630a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 532.680741] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d944ea62-cb8e-4a54-8a16-81477a0272d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.690417] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01c4eba-5729-4ad6-83a9-e178024e4771 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.716976] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf could not be found. [ 532.717386] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.718897] env[63175]: INFO nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 532.718897] env[63175]: DEBUG oslo.service.loopingcall [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.718897] env[63175]: DEBUG nova.compute.manager [-] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 532.718897] env[63175]: DEBUG nova.network.neutron [-] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.744849] env[63175]: DEBUG nova.network.neutron [-] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.806288] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 532.839486] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.839732] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.839886] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.840086] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.840234] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.840381] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.840583] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.840739] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.840927] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.841123] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.841300] env[63175]: DEBUG nova.virt.hardware [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.842152] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8adc2db-1c71-4050-8e3a-2f88d92a2a05 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.850708] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db25461-37d1-4825-9200-ff7584f3981d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.864370] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 532.871124] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Creating folder: Project (e4b8d70bd69a4f7a892a12c1ddbfccc4). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 532.871450] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f8979fe-fb5d-4711-8e97-e3581ec2672a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.881887] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247839, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.884153] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Created folder: Project (e4b8d70bd69a4f7a892a12c1ddbfccc4) in parent group-v268956. [ 532.884153] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Creating folder: Instances. Parent ref: group-v268960. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 532.884153] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd5955a0-3a61-424a-8f3c-ede57f164103 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.891783] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Created folder: Instances in parent group-v268960. [ 532.892071] env[63175]: DEBUG oslo.service.loopingcall [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.892227] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 532.892769] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e314602-b021-4357-9bd6-448a1821a241 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.910450] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 532.910450] env[63175]: value = "task-1247842" [ 532.910450] env[63175]: _type = "Task" [ 532.910450] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.917911] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.161941] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.175615] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 533.175615] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.256s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.175615] env[63175]: INFO nova.compute.claims [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.221241] env[63175]: DEBUG nova.network.neutron [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.247809] env[63175]: DEBUG nova.network.neutron [-] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.348731] env[63175]: DEBUG nova.network.neutron [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.387707] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247839, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.48756} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.388473] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 533.388473] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 533.388473] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79e17fc4-9f89-4014-b4d7-ebb5e2e0143c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.398154] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 533.398154] env[63175]: value = "task-1247843" [ 533.398154] env[63175]: _type = "Task" [ 533.398154] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.410301] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.426954] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.476186] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "e96543c8-dc4e-45b3-8ece-297ee6dbcedb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.480877] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "e96543c8-dc4e-45b3-8ece-297ee6dbcedb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.685209] env[63175]: DEBUG nova.compute.utils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.687526] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 533.687710] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 533.751297] env[63175]: INFO nova.compute.manager [-] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Took 1.03 seconds to deallocate network for instance. [ 533.755778] env[63175]: DEBUG nova.compute.claims [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.755778] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.787857] env[63175]: ERROR nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 533.787857] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 533.787857] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.787857] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.787857] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.787857] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.787857] env[63175]: ERROR nova.compute.manager raise self.value [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.787857] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.787857] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.787857] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.788629] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.788629] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.788629] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 533.788629] env[63175]: ERROR nova.compute.manager [ 533.788629] env[63175]: Traceback (most recent call last): [ 533.788849] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.788849] env[63175]: listener.cb(fileno) [ 533.788849] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.788849] env[63175]: result = function(*args, **kwargs) [ 533.788849] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.788849] env[63175]: return func(*args, **kwargs) [ 533.788849] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 533.788849] env[63175]: raise e [ 533.788849] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 533.788849] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 533.788849] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.788849] env[63175]: created_port_ids = self._update_ports_for_instance( [ 533.788849] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.788849] env[63175]: with excutils.save_and_reraise_exception(): [ 533.788849] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.788849] env[63175]: self.force_reraise() [ 533.788849] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.788849] env[63175]: raise self.value [ 533.788849] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.788849] env[63175]: updated_port = self._update_port( [ 533.788849] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.788849] env[63175]: _ensure_no_port_binding_failure(port) [ 533.788849] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.788849] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.788849] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 533.788849] env[63175]: Removing descriptor: 15 [ 533.790474] env[63175]: ERROR nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Traceback (most recent call last): [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] yield resources [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.driver.spawn(context, instance, image_meta, [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] vm_ref = self.build_virtual_machine(instance, [ 533.790474] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] for vif in network_info: [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self._sync_wrapper(fn, *args, **kwargs) [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.wait() [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self[:] = self._gt.wait() [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self._exit_event.wait() [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.790976] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] result = hub.switch() [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self.greenlet.switch() [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] result = function(*args, **kwargs) [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return func(*args, **kwargs) [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise e [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] nwinfo = self.network_api.allocate_for_instance( [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] created_port_ids = self._update_ports_for_instance( [ 533.791440] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] with excutils.save_and_reraise_exception(): [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.force_reraise() [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise self.value [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] updated_port = self._update_port( [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] _ensure_no_port_binding_failure(port) [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise exception.PortBindingFailed(port_id=port['id']) [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 533.791855] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] [ 533.792538] env[63175]: INFO nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Terminating instance [ 533.821933] env[63175]: DEBUG nova.policy [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afe7974a141b401a82f7f352825fddb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '171e1987d09940dcb67643a6b7bbf2b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.850617] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a062207-d7cb-4a61-badc-a13c6d3f1f77 req-e206a949-918d-4d6e-9a7b-3382f418699e service nova] Releasing lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.907732] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068174} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.908039] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 533.908898] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d1b693-95f5-46f0-a83b-629d6708c52e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.935508] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 533.935508] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93e2cce5-6f30-43cf-9f17-2267812090f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.953486] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.959927] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 533.959927] env[63175]: value = "task-1247844" [ 533.959927] env[63175]: _type = "Task" [ 533.959927] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.967745] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247844, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.188753] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 534.301153] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquiring lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.301153] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquired lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.301153] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.407737] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1df8e79-d411-4d38-b1fb-8df5fcab8be2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.420431] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411ec524-25a2-4ca8-ae58-7139e97048ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.428152] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.453525] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dff03c-ed61-4d11-a4e2-2c86ee312c9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.466435] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65182432-31b6-4138-9ddb-7fa21bd6b3a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.474777] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247844, 'name': ReconfigVM_Task, 'duration_secs': 0.27256} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.483939] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 534.483939] env[63175]: DEBUG nova.compute.provider_tree [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.485106] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42dcce2b-fd21-4645-b422-59efc24bec8d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.491653] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 534.491653] env[63175]: value = "task-1247845" [ 534.491653] env[63175]: _type = "Task" [ 534.491653] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.500928] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247845, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.821411] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.829791] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquiring lock "f33d2617-c4ef-4174-a838-ce0558867b9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.830049] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "f33d2617-c4ef-4174-a838-ce0558867b9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.890900] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.931261] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.989012] env[63175]: DEBUG nova.scheduler.client.report [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 535.006933] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247845, 'name': Rename_Task, 'duration_secs': 0.14483} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.007048] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 535.007333] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35b11fbe-2ef4-4886-8bc8-18040adae19a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.015186] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 535.015186] env[63175]: value = "task-1247846" [ 535.015186] env[63175]: _type = "Task" [ 535.015186] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.025531] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.026803] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Successfully created port: 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.174725] env[63175]: DEBUG nova.compute.manager [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Received event network-changed-0d58a188-6ddc-4364-9486-421c86294015 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 535.175094] env[63175]: DEBUG nova.compute.manager [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Refreshing instance network info cache due to event network-changed-0d58a188-6ddc-4364-9486-421c86294015. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 535.175481] env[63175]: DEBUG oslo_concurrency.lockutils [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] Acquiring lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.211490] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 535.239868] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 535.240386] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 535.240652] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 535.241050] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 535.241522] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 535.241757] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 535.242042] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 535.242406] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 535.242712] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 535.243151] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 535.243443] env[63175]: DEBUG nova.virt.hardware [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 535.244830] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4728fe-d624-4e9e-b395-e7f1103e3a84 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.254907] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180c2f05-802f-489d-95bc-e8867e7696b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.397022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Releasing lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.397022] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 535.397022] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.397022] env[63175]: DEBUG oslo_concurrency.lockutils [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] Acquired lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.397022] env[63175]: DEBUG nova.network.neutron [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Refreshing network info cache for port 0d58a188-6ddc-4364-9486-421c86294015 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 535.397354] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-026620c1-7afa-45a0-9d28-b03b049e4ad1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.410019] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f3d8db-1f10-4552-bacb-c05469a70105 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.432690] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.439198] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 18acf9cd-0661-48c0-8fd2-b80edba4dfa6 could not be found. [ 535.440207] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.440207] env[63175]: INFO nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 535.440207] env[63175]: DEBUG oslo.service.loopingcall [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.440623] env[63175]: DEBUG nova.compute.manager [-] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 535.440795] env[63175]: DEBUG nova.network.neutron [-] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.461799] env[63175]: DEBUG nova.network.neutron [-] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.501763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.502283] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 535.505488] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.011s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.506980] env[63175]: INFO nova.compute.claims [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.524601] env[63175]: DEBUG oslo_vmware.api [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247846, 'name': PowerOnVM_Task, 'duration_secs': 0.48593} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.524769] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 535.525983] env[63175]: INFO nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Took 9.60 seconds to spawn the instance on the hypervisor. [ 535.525983] env[63175]: DEBUG nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 535.526399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2461e78d-abf5-45d2-b47c-123f2a4d1411 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.821765] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquiring lock "edd63563-8fe3-4280-9300-8a11ba40ba68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.822033] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "edd63563-8fe3-4280-9300-8a11ba40ba68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.933455] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247842, 'name': CreateVM_Task, 'duration_secs': 2.943506} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.933650] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 535.936410] env[63175]: DEBUG oslo_vmware.service [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e93a5c3-1f82-4f3e-a2bd-6faa0c7dcb7f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.946433] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.946600] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.947051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 535.947314] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f62f1be9-3a90-4048-a3ff-4483ceff9ed7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.955904] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 535.955904] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526324b6-b6b6-4226-b65e-c408ce1ced75" [ 535.955904] env[63175]: _type = "Task" [ 535.955904] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.960306] env[63175]: DEBUG nova.network.neutron [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.964484] env[63175]: DEBUG nova.network.neutron [-] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.970694] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526324b6-b6b6-4226-b65e-c408ce1ced75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.008029] env[63175]: DEBUG nova.compute.utils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.009200] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 536.009504] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 536.056174] env[63175]: INFO nova.compute.manager [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Took 15.63 seconds to build instance. [ 536.059150] env[63175]: DEBUG nova.compute.manager [req-8be33229-ba79-4a7e-bd32-bf444c6ea3e0 req-cad9d707-0a83-4c4b-a1f8-2f6d7ee6c04a service nova] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Received event network-vif-deleted-edfa6f97-302b-41e5-8c84-a3add567630a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 536.107371] env[63175]: DEBUG nova.network.neutron [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.241549] env[63175]: DEBUG nova.policy [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '49765f8191104bfe8965327e32f30e34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9215ef450b174466a5d28101b9314e7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.466310] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.466605] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 536.466831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.466984] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.467193] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 536.467439] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af2e1e1f-04a5-4059-b7ca-53d49645b80a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.469441] env[63175]: INFO nova.compute.manager [-] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Took 1.03 seconds to deallocate network for instance. [ 536.471637] env[63175]: DEBUG nova.compute.claims [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.471808] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.486678] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 536.486898] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 536.487710] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f6c655-2ba9-4451-8f77-63f5addac8cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.498713] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-683ebcc1-2c05-495d-8a27-ba9018a3d76d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.508111] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 536.508111] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52012444-5b2b-98ed-0bf3-018b35436e27" [ 536.508111] env[63175]: _type = "Task" [ 536.508111] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.514126] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 536.521961] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52012444-5b2b-98ed-0bf3-018b35436e27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.560482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-57db82e5-5c9d-498c-9cb6-c15a1c8676f4 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.144s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.609073] env[63175]: DEBUG oslo_concurrency.lockutils [req-aa2e2104-c5e6-4674-8be6-a9c54422751c req-f28ca042-0553-4c5d-8b03-8f21c2f5adb8 service nova] Releasing lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.746981] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c28a1e6-5466-46c3-80db-29d68db38bac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.757984] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e0bd21-646f-4810-80b6-73f418862d43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.794676] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5421959e-5591-4329-aa44-f47451b7fa5e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.805344] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d206d5-4be0-4420-b851-53873688cd79 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.820863] env[63175]: DEBUG nova.compute.provider_tree [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.021714] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Preparing fetch location {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 537.022018] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Creating directory with path [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 537.022198] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15570ed1-ad54-4080-baaf-ae5eec347ed9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.048549] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Created directory with path [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 537.049239] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Fetch image to [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 537.049239] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Downloading image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk on the data store datastore1 {{(pid=63175) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 537.049921] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b6a014-415f-4663-97a8-eb6c6ebbec3f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.061399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd275a94-68ac-4c24-9086-aace268eca4b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.070885] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 537.075341] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c055ba-8c6f-4cbd-9e80-92d3ea6c72e3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.112912] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a8d5c7-38f0-4df9-a09c-12921f2e105d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.120731] env[63175]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c818f3be-da3e-409c-b302-6862e8aac89e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.145314] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Downloading image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to the data store datastore1 {{(pid=63175) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 537.206378] env[63175]: DEBUG oslo_vmware.rw_handles [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63175) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 537.328772] env[63175]: DEBUG nova.scheduler.client.report [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 537.529306] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 537.559323] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.559555] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.559707] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.559929] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.560091] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.560258] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.560465] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.560586] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.560743] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.560896] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.561171] env[63175]: DEBUG nova.virt.hardware [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.562103] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab418a60-b44d-442e-8d21-ca9b87c2349b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.575890] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed73cad-31d5-4c48-aacc-17ba0a54d69f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.619300] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.664372] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Successfully created port: 231db14b-d441-4dc0-aaa3-3534e92fb7e4 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.832608] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.833570] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 537.837094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.073s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.839415] env[63175]: INFO nova.compute.claims [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.924342] env[63175]: DEBUG oslo_vmware.rw_handles [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Completed reading data from the image iterator. {{(pid=63175) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 537.924669] env[63175]: DEBUG oslo_vmware.rw_handles [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 537.957605] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquiring lock "16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.958100] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.072962] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Downloaded image file data 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk on the data store datastore1 {{(pid=63175) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 538.074662] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Caching image {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 538.074913] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Copying Virtual Disk [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk to [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 538.075221] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14cdcedb-ab1e-42a0-b5aa-9b77618ba8cf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.084118] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 538.084118] env[63175]: value = "task-1247847" [ 538.084118] env[63175]: _type = "Task" [ 538.084118] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.093040] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.255447] env[63175]: DEBUG nova.compute.manager [req-042bfe21-aad2-4859-a736-92ce0bb023c1 req-9df769a9-de69-4e36-b887-e9c9b587f12f service nova] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Received event network-vif-deleted-0d58a188-6ddc-4364-9486-421c86294015 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 538.258641] env[63175]: INFO nova.compute.manager [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Rebuilding instance [ 538.310724] env[63175]: DEBUG nova.compute.manager [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 538.311898] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b21508b-ab30-49dc-a451-05ec0c87f948 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.347189] env[63175]: DEBUG nova.compute.utils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.350036] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 538.351081] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 538.443133] env[63175]: DEBUG nova.policy [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18b70e10722c4a47907ce844dcf41ca2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bff8e7afe6d4bcc98b46f4dd8becff7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.596564] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247847, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.855680] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 538.989608] env[63175]: ERROR nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 538.989608] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 538.989608] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.989608] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.989608] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.989608] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.989608] env[63175]: ERROR nova.compute.manager raise self.value [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.989608] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.989608] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.989608] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.990068] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.990068] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.990068] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 538.990068] env[63175]: ERROR nova.compute.manager [ 538.990068] env[63175]: Traceback (most recent call last): [ 538.990068] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.990068] env[63175]: listener.cb(fileno) [ 538.990068] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.990068] env[63175]: result = function(*args, **kwargs) [ 538.990068] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.990068] env[63175]: return func(*args, **kwargs) [ 538.990068] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 538.990068] env[63175]: raise e [ 538.990068] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 538.990068] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 538.990068] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.990068] env[63175]: created_port_ids = self._update_ports_for_instance( [ 538.990068] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.990068] env[63175]: with excutils.save_and_reraise_exception(): [ 538.990068] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.990068] env[63175]: self.force_reraise() [ 538.990068] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.990068] env[63175]: raise self.value [ 538.990068] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.990068] env[63175]: updated_port = self._update_port( [ 538.990068] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.990068] env[63175]: _ensure_no_port_binding_failure(port) [ 538.990068] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.990068] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.990999] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 538.990999] env[63175]: Removing descriptor: 17 [ 538.990999] env[63175]: ERROR nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Traceback (most recent call last): [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] yield resources [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.driver.spawn(context, instance, image_meta, [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.990999] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] vm_ref = self.build_virtual_machine(instance, [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] for vif in network_info: [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self._sync_wrapper(fn, *args, **kwargs) [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.wait() [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self[:] = self._gt.wait() [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self._exit_event.wait() [ 538.991344] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] result = hub.switch() [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self.greenlet.switch() [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] result = function(*args, **kwargs) [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return func(*args, **kwargs) [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise e [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] nwinfo = self.network_api.allocate_for_instance( [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.991644] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] created_port_ids = self._update_ports_for_instance( [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] with excutils.save_and_reraise_exception(): [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.force_reraise() [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise self.value [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] updated_port = self._update_port( [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] _ensure_no_port_binding_failure(port) [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.991963] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise exception.PortBindingFailed(port_id=port['id']) [ 538.993613] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 538.993613] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] [ 538.993613] env[63175]: INFO nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Terminating instance [ 539.028940] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Successfully created port: d59a71f1-94b2-4955-bd65-ae6e532707ae {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.096542] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649947} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.096542] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Copied Virtual Disk [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk to [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 539.097834] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleting the datastore file [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/tmp-sparse.vmdk {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 539.097834] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-acd7fc53-fc2d-4e4b-a32d-220e276e59c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.106614] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 539.106614] env[63175]: value = "task-1247848" [ 539.106614] env[63175]: _type = "Task" [ 539.106614] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.122100] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.141666] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44065d45-b641-4f33-838f-5d53130ae263 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.148311] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86edb23-f1e6-4492-8abe-afbcbaba56b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.182864] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7955608-b61f-42e9-aad6-0be804a351a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.191046] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac967dd-db17-46fa-9209-7aef23f5a32d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.206092] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 539.330273] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 539.330273] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-253da717-28f1-4c9a-8a48-2e89ab66ea7f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.339407] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 539.339407] env[63175]: value = "task-1247849" [ 539.339407] env[63175]: _type = "Task" [ 539.339407] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.346501] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.499147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.499341] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquired lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.499548] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.621511] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.058393} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.622145] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 539.622145] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Moving file from [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa/8e9c4b2c-a2a7-4856-ba6d-ed187f088508 to [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508. {{(pid=63175) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 539.622145] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c64fb409-94f0-4390-aa81-f08767d05244 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.631483] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 539.631483] env[63175]: value = "task-1247850" [ 539.631483] env[63175]: _type = "Task" [ 539.631483] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.644944] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247850, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.728285] env[63175]: ERROR nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [req-ba2bc6fd-84ce-453a-ba67-3f6e503c1a3d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f10bd0b1-85ed-4db3-a7f2-6e16e785219e. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ba2bc6fd-84ce-453a-ba67-3f6e503c1a3d"}]} [ 539.752182] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Refreshing inventories for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 539.771132] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating ProviderTree inventory for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 539.771457] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 539.786324] env[63175]: DEBUG nova.compute.manager [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Received event network-changed-4d86bf78-20e6-4fd7-bdcc-6fe89704e13d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 539.786324] env[63175]: DEBUG nova.compute.manager [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Refreshing instance network info cache due to event network-changed-4d86bf78-20e6-4fd7-bdcc-6fe89704e13d. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 539.786510] env[63175]: DEBUG oslo_concurrency.lockutils [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] Acquiring lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.790464] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Refreshing aggregate associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, aggregates: None {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 539.816468] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Refreshing trait associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 539.849317] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247849, 'name': PowerOffVM_Task, 'duration_secs': 0.118796} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.849600] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 539.849844] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.850634] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fa6f68-1e3a-4072-ae10-205d81937877 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.859655] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 539.860292] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-411b38c8-e110-4982-a6b2-53bc929ea22f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.869492] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 539.888909] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 539.889292] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 539.889292] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleting the datastore file [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 539.891675] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84446899-cdb3-4849-8bd3-49d315bc23db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.901186] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 539.901186] env[63175]: value = "task-1247852" [ 539.901186] env[63175]: _type = "Task" [ 539.901186] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.904203] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.904473] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.904625] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.904805] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.904954] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.905109] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.905309] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.905461] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.905717] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.907015] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.907015] env[63175]: DEBUG nova.virt.hardware [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.909822] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f44f967-6763-4206-bf4b-69a5b3346c23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.922382] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.922781] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30c22d9-d8c4-41b6-a442-b158290ff5ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.030978] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.082330] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a76f6d-5d15-41b7-8aab-96ec94b7f1c6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.090806] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecba81fc-f662-464b-b3d4-f329a4b21958 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.125818] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.128163] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf530741-c9d6-419f-806c-803a3d199a67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.142540] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6b4815-6111-4d1a-b6ec-b35214ae4300 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.149921] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247850, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024521} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.150603] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] File moved {{(pid=63175) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 540.150842] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Cleaning up location [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 540.151179] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleting the datastore file [datastore1] vmware_temp/87af3694-a5aa-4442-9a57-6276798052fa {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 540.151948] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85651d5f-4d18-4f4e-ad12-a63c0b3be48f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.165099] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 540.171258] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 540.171258] env[63175]: value = "task-1247853" [ 540.171258] env[63175]: _type = "Task" [ 540.171258] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.181381] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247853, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.280228] env[63175]: ERROR nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 540.280228] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.280228] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.280228] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.280228] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.280228] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.280228] env[63175]: ERROR nova.compute.manager raise self.value [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.280228] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.280228] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.280228] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.280966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.280966] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.280966] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 540.280966] env[63175]: ERROR nova.compute.manager [ 540.280966] env[63175]: Traceback (most recent call last): [ 540.280966] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.280966] env[63175]: listener.cb(fileno) [ 540.280966] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.280966] env[63175]: result = function(*args, **kwargs) [ 540.280966] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.280966] env[63175]: return func(*args, **kwargs) [ 540.280966] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.280966] env[63175]: raise e [ 540.280966] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.280966] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 540.280966] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.280966] env[63175]: created_port_ids = self._update_ports_for_instance( [ 540.280966] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.280966] env[63175]: with excutils.save_and_reraise_exception(): [ 540.280966] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.280966] env[63175]: self.force_reraise() [ 540.280966] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.280966] env[63175]: raise self.value [ 540.280966] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.280966] env[63175]: updated_port = self._update_port( [ 540.280966] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.280966] env[63175]: _ensure_no_port_binding_failure(port) [ 540.280966] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.280966] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.281748] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 540.281748] env[63175]: Removing descriptor: 16 [ 540.282998] env[63175]: ERROR nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] Traceback (most recent call last): [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] yield resources [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.driver.spawn(context, instance, image_meta, [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] vm_ref = self.build_virtual_machine(instance, [ 540.282998] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] for vif in network_info: [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self._sync_wrapper(fn, *args, **kwargs) [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.wait() [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self[:] = self._gt.wait() [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self._exit_event.wait() [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.284263] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] result = hub.switch() [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self.greenlet.switch() [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] result = function(*args, **kwargs) [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return func(*args, **kwargs) [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise e [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] nwinfo = self.network_api.allocate_for_instance( [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] created_port_ids = self._update_ports_for_instance( [ 540.284613] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] with excutils.save_and_reraise_exception(): [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.force_reraise() [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise self.value [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] updated_port = self._update_port( [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] _ensure_no_port_binding_failure(port) [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise exception.PortBindingFailed(port_id=port['id']) [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 540.284961] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] [ 540.285290] env[63175]: INFO nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Terminating instance [ 540.410818] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098572} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.412018] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.412278] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 540.412490] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.632841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Releasing lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.633296] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 540.633495] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.633838] env[63175]: DEBUG oslo_concurrency.lockutils [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] Acquired lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.633966] env[63175]: DEBUG nova.network.neutron [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Refreshing network info cache for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 540.635043] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e091fba8-8faf-4ffa-9a1d-604d5c11f2b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.644385] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cb2402-ffe4-458d-bdac-1c5416f73f99 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.671540] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 380a6278-22d4-485b-b9cb-ccb1ac07d4a9 could not be found. [ 540.671540] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.671633] env[63175]: INFO nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 540.672291] env[63175]: DEBUG oslo.service.loopingcall [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.672708] env[63175]: DEBUG nova.compute.manager [-] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 540.672708] env[63175]: DEBUG nova.network.neutron [-] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.684603] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247853, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026684} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.684963] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.685877] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d42bb3e6-c8a3-4313-b433-28a64c8828d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.692558] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 540.692558] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dfaa3e-f1ef-9e85-9c7e-bae5906e97a9" [ 540.692558] env[63175]: _type = "Task" [ 540.692558] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.702057] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dfaa3e-f1ef-9e85-9c7e-bae5906e97a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.707066] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updated inventory for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with generation 18 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 540.707288] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e generation from 18 to 19 during operation: update_inventory {{(pid=63175) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 540.707460] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 540.711733] env[63175]: DEBUG nova.network.neutron [-] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.788302] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquiring lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.788899] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquired lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.789326] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.101185] env[63175]: ERROR nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 541.101185] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 541.101185] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.101185] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.101185] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.101185] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.101185] env[63175]: ERROR nova.compute.manager raise self.value [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.101185] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.101185] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.101185] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.101698] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.101698] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.101698] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 541.101698] env[63175]: ERROR nova.compute.manager [ 541.101698] env[63175]: Traceback (most recent call last): [ 541.101698] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.101698] env[63175]: listener.cb(fileno) [ 541.101698] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.101698] env[63175]: result = function(*args, **kwargs) [ 541.101698] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.101698] env[63175]: return func(*args, **kwargs) [ 541.101698] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 541.101698] env[63175]: raise e [ 541.101698] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 541.101698] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 541.101698] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.101698] env[63175]: created_port_ids = self._update_ports_for_instance( [ 541.101698] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.101698] env[63175]: with excutils.save_and_reraise_exception(): [ 541.101698] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.101698] env[63175]: self.force_reraise() [ 541.101698] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.101698] env[63175]: raise self.value [ 541.101698] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.101698] env[63175]: updated_port = self._update_port( [ 541.101698] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.101698] env[63175]: _ensure_no_port_binding_failure(port) [ 541.101698] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.101698] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.102645] env[63175]: nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 541.102645] env[63175]: Removing descriptor: 15 [ 541.102645] env[63175]: ERROR nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Traceback (most recent call last): [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] yield resources [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.driver.spawn(context, instance, image_meta, [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.102645] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] vm_ref = self.build_virtual_machine(instance, [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] for vif in network_info: [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self._sync_wrapper(fn, *args, **kwargs) [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.wait() [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self[:] = self._gt.wait() [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self._exit_event.wait() [ 541.103038] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] result = hub.switch() [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self.greenlet.switch() [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] result = function(*args, **kwargs) [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return func(*args, **kwargs) [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise e [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] nwinfo = self.network_api.allocate_for_instance( [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.103384] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] created_port_ids = self._update_ports_for_instance( [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] with excutils.save_and_reraise_exception(): [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.force_reraise() [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise self.value [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] updated_port = self._update_port( [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] _ensure_no_port_binding_failure(port) [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.103703] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise exception.PortBindingFailed(port_id=port['id']) [ 541.103992] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 541.103992] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] [ 541.103992] env[63175]: INFO nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Terminating instance [ 541.164062] env[63175]: DEBUG nova.network.neutron [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.206379] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dfaa3e-f1ef-9e85-9c7e-bae5906e97a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009312} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.206565] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.206803] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 56b3ba41-0406-4b69-89af-b00e87feaa89/56b3ba41-0406-4b69-89af-b00e87feaa89.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 541.207194] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd9a3ea2-ef73-4a26-afc1-5ea086766c3c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.215021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.378s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.215510] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 541.218074] env[63175]: DEBUG nova.network.neutron [-] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.219204] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 541.219204] env[63175]: value = "task-1247854" [ 541.219204] env[63175]: _type = "Task" [ 541.219204] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.219667] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.289s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.232174] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.275475] env[63175]: DEBUG nova.network.neutron [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.322629] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.453871] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.464469] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.464772] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.464969] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.465175] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.467023] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.467023] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.467023] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.467023] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.467023] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.467291] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.467291] env[63175]: DEBUG nova.virt.hardware [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.468331] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a81b73a-c7dd-4f66-a5eb-4e2343c03bed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.481286] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1299bee6-c072-4c96-9292-87442166d7b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.500443] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 541.504166] env[63175]: DEBUG oslo.service.loopingcall [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.504533] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 541.504817] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03381620-0fe8-4f77-b839-e6e0a9c7a905 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.526864] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 541.526864] env[63175]: value = "task-1247855" [ 541.526864] env[63175]: _type = "Task" [ 541.526864] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.535514] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247855, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.612909] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquiring lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.612909] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquired lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.612909] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.722593] env[63175]: DEBUG nova.compute.utils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.724511] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 541.724600] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 541.729345] env[63175]: INFO nova.compute.manager [-] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Took 1.06 seconds to deallocate network for instance. [ 541.741973] env[63175]: DEBUG nova.compute.claims [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.742238] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.750669] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468472} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.750669] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 56b3ba41-0406-4b69-89af-b00e87feaa89/56b3ba41-0406-4b69-89af-b00e87feaa89.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 541.750669] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 541.750669] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ebca7d4-0671-4c92-a9ec-8b748fd7cded {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.757434] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 541.757434] env[63175]: value = "task-1247856" [ 541.757434] env[63175]: _type = "Task" [ 541.757434] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.772997] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.778145] env[63175]: DEBUG oslo_concurrency.lockutils [req-ab4611ef-f450-487e-b994-ebb479fcc411 req-4fdaeaaf-a5de-4fdb-9ce2-809da8d0d75b service nova] Releasing lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.876187] env[63175]: DEBUG nova.policy [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f79e0b1c8f2c427cb83e71501e0f9b30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '265d6942195e499cbd44d67393cc8ee5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.961835] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Releasing lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.961953] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 541.963016] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.963016] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30ffa5f8-7092-4d6d-92c1-4db00f9aa454 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.972574] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3a69f1-9b9b-4372-af3a-68dd513fdd13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.985808] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e1bf8f-7919-44dc-8151-8786c1010a5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.994082] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc109b2-838c-4273-a86f-8e4f0b78d485 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.002997] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1561f155-ff79-4790-a164-43f08230e229 could not be found. [ 542.003245] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.003424] env[63175]: INFO nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.005140] env[63175]: DEBUG oslo.service.loopingcall [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.005140] env[63175]: DEBUG nova.compute.manager [-] [instance: 1561f155-ff79-4790-a164-43f08230e229] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 542.005140] env[63175]: DEBUG nova.network.neutron [-] [instance: 1561f155-ff79-4790-a164-43f08230e229] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.033492] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa80514-c990-488a-99ba-26e7eeb43c05 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.043627] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247855, 'name': CreateVM_Task, 'duration_secs': 0.309877} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.045918] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 542.046012] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.046144] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.046609] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 542.047652] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f17daa-96d6-43b4-a83c-c972c2a8f153 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.051458] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4da7f5f5-a319-46a6-b9de-ec514355a419 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.056083] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 542.056083] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521bf004-3586-19c8-2b4b-e9a102bdc16b" [ 542.056083] env[63175]: _type = "Task" [ 542.056083] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.064417] env[63175]: DEBUG nova.compute.provider_tree [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.073895] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521bf004-3586-19c8-2b4b-e9a102bdc16b, 'name': SearchDatastore_Task, 'duration_secs': 0.009366} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.074066] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.074324] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 542.074514] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.074659] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.074832] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 542.075714] env[63175]: DEBUG nova.network.neutron [-] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.079036] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfd8dbb5-a6fa-4419-be0c-fc4b7e8ac717 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.090112] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 542.090269] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 542.091204] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b539fd02-c389-46db-b2ce-53b161600e5b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.096321] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 542.096321] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208d5ac-c231-9a0a-c18b-f4793f57b374" [ 542.096321] env[63175]: _type = "Task" [ 542.096321] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.103858] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208d5ac-c231-9a0a-c18b-f4793f57b374, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.231330] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 542.267158] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060013} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.267309] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 542.269546] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f9ede7-b6df-4fc8-b7e2-d2dd7126ba95 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.288030] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] 56b3ba41-0406-4b69-89af-b00e87feaa89/56b3ba41-0406-4b69-89af-b00e87feaa89.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 542.288270] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.290123] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4b0a89a-5eda-47ce-bedd-ddd4fe705e8b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.310361] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 542.310361] env[63175]: value = "task-1247857" [ 542.310361] env[63175]: _type = "Task" [ 542.310361] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.318432] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247857, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.438724] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Successfully created port: c79ee925-f64d-4b19-b0c7-d4b8214c72f9 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.449164] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.569707] env[63175]: DEBUG nova.scheduler.client.report [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 542.582643] env[63175]: DEBUG nova.network.neutron [-] [instance: 1561f155-ff79-4790-a164-43f08230e229] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.608252] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208d5ac-c231-9a0a-c18b-f4793f57b374, 'name': SearchDatastore_Task, 'duration_secs': 0.008043} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.609089] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68fa5e72-4a49-4280-8878-8dcf124a0056 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.614907] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 542.614907] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52651a19-92bf-bf51-8abf-baeb20e94ec3" [ 542.614907] env[63175]: _type = "Task" [ 542.614907] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.624506] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52651a19-92bf-bf51-8abf-baeb20e94ec3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.783217] env[63175]: DEBUG nova.compute.manager [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Received event network-changed-d59a71f1-94b2-4955-bd65-ae6e532707ae {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 542.783217] env[63175]: DEBUG nova.compute.manager [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Refreshing instance network info cache due to event network-changed-d59a71f1-94b2-4955-bd65-ae6e532707ae. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 542.783821] env[63175]: DEBUG oslo_concurrency.lockutils [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] Acquiring lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.823732] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247857, 'name': ReconfigVM_Task, 'duration_secs': 0.257841} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.824088] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Reconfigured VM instance instance-00000005 to attach disk [datastore1] 56b3ba41-0406-4b69-89af-b00e87feaa89/56b3ba41-0406-4b69-89af-b00e87feaa89.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 542.828632] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c76ccbfa-744c-4ee7-af1d-13a463f22755 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.836752] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 542.836752] env[63175]: value = "task-1247858" [ 542.836752] env[63175]: _type = "Task" [ 542.836752] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.844856] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247858, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.952670] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Releasing lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.953179] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 542.953531] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.953745] env[63175]: DEBUG oslo_concurrency.lockutils [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] Acquired lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.954048] env[63175]: DEBUG nova.network.neutron [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Refreshing network info cache for port d59a71f1-94b2-4955-bd65-ae6e532707ae {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.956133] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b0d19af-4d71-4fd9-9078-4a7370771a5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.969189] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b52e32-5e5e-4222-a9a3-9b5471f46e2d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.992548] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0e0f0535-d539-4572-8752-e3e6a4b79b8c could not be found. [ 542.992797] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.992980] env[63175]: INFO nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.993240] env[63175]: DEBUG oslo.service.loopingcall [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.993840] env[63175]: DEBUG nova.compute.manager [-] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 542.993840] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 543.017299] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.074909] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.075685] env[63175]: ERROR nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Traceback (most recent call last): [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.driver.spawn(context, instance, image_meta, [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] vm_ref = self.build_virtual_machine(instance, [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.075685] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] for vif in network_info: [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self._sync_wrapper(fn, *args, **kwargs) [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.wait() [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self[:] = self._gt.wait() [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self._exit_event.wait() [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] result = hub.switch() [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.076122] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return self.greenlet.switch() [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] result = function(*args, **kwargs) [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] return func(*args, **kwargs) [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise e [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] nwinfo = self.network_api.allocate_for_instance( [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] created_port_ids = self._update_ports_for_instance( [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] with excutils.save_and_reraise_exception(): [ 543.076873] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] self.force_reraise() [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise self.value [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] updated_port = self._update_port( [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] _ensure_no_port_binding_failure(port) [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] raise exception.PortBindingFailed(port_id=port['id']) [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] nova.exception.PortBindingFailed: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. [ 543.077350] env[63175]: ERROR nova.compute.manager [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] [ 543.077679] env[63175]: DEBUG nova.compute.utils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.077679] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.400s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.079165] env[63175]: INFO nova.compute.claims [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.084608] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Build of instance 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52 was re-scheduled: Binding failed for port e4fa27c5-f657-463e-b046-8fc23264bd9d, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 543.087119] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 543.087363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquiring lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.088598] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Acquired lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.088598] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.092881] env[63175]: INFO nova.compute.manager [-] [instance: 1561f155-ff79-4790-a164-43f08230e229] Took 1.09 seconds to deallocate network for instance. [ 543.094551] env[63175]: DEBUG nova.compute.claims [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.094551] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.126403] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52651a19-92bf-bf51-8abf-baeb20e94ec3, 'name': SearchDatastore_Task, 'duration_secs': 0.008979} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.126857] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.127031] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 543.127321] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01802a5e-4e66-4d95-bb2f-75a8396c6919 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.134195] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 543.134195] env[63175]: value = "task-1247859" [ 543.134195] env[63175]: _type = "Task" [ 543.134195] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.144105] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247859, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.244903] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 543.281408] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.281554] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.281601] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.283172] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.283172] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.283172] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.283172] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.283172] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.283326] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.283326] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.283425] env[63175]: DEBUG nova.virt.hardware [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.284324] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff01a52c-0329-4371-a236-fa251a68e2b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.292452] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b5b672-7f9e-4adc-9036-88e42f9e9215 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.347565] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247858, 'name': Rename_Task, 'duration_secs': 0.143294} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.348085] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 543.348507] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72887e74-3277-497b-b61d-58f4bdff8e7e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.359257] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 543.359257] env[63175]: value = "task-1247860" [ 543.359257] env[63175]: _type = "Task" [ 543.359257] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.368952] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.489035] env[63175]: DEBUG nova.network.neutron [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.521046] env[63175]: DEBUG nova.network.neutron [-] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.643283] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.653376] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247859, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.751571] env[63175]: DEBUG nova.network.neutron [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.813882] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.871813] env[63175]: DEBUG oslo_vmware.api [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247860, 'name': PowerOnVM_Task, 'duration_secs': 0.49009} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.872424] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 543.872725] env[63175]: INFO nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Took 11.07 seconds to spawn the instance on the hypervisor. [ 543.872937] env[63175]: DEBUG nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 543.873716] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a56352-b868-457e-b291-28bff82d391e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.895067] env[63175]: DEBUG nova.compute.manager [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Received event network-vif-deleted-4d86bf78-20e6-4fd7-bdcc-6fe89704e13d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 543.895299] env[63175]: DEBUG nova.compute.manager [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Received event network-changed-231db14b-d441-4dc0-aaa3-3534e92fb7e4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 543.895478] env[63175]: DEBUG nova.compute.manager [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Refreshing instance network info cache due to event network-changed-231db14b-d441-4dc0-aaa3-3534e92fb7e4. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 543.895739] env[63175]: DEBUG oslo_concurrency.lockutils [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] Acquiring lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.895901] env[63175]: DEBUG oslo_concurrency.lockutils [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] Acquired lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.896009] env[63175]: DEBUG nova.network.neutron [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Refreshing network info cache for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.023733] env[63175]: INFO nova.compute.manager [-] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Took 1.03 seconds to deallocate network for instance. [ 544.026076] env[63175]: DEBUG nova.compute.claims [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 544.026258] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.147834] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247859, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521286} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.148133] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 544.148346] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 544.148593] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e195d04-2323-49c9-acf0-ce4743451657 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.155098] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 544.155098] env[63175]: value = "task-1247861" [ 544.155098] env[63175]: _type = "Task" [ 544.155098] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.168511] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.254812] env[63175]: DEBUG oslo_concurrency.lockutils [req-777ae4a3-a662-4c2c-ad40-8c44e78c1bed req-88c2e1e8-9eb5-409c-9521-db08780a803b service nova] Releasing lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.319049] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Releasing lock "refresh_cache-0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.319298] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 544.319470] env[63175]: DEBUG nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 544.319650] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.347448] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.353052] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b65912-57b0-483c-a256-15011f1ab069 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.367227] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee826935-b306-4632-980a-ed9ba5f3c1e1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.417166] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bafa718-7083-423e-a424-06cb29355d3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.424178] env[63175]: INFO nova.compute.manager [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Took 22.36 seconds to build instance. [ 544.431459] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eecfdbd-7a76-462d-ade4-8fd80977c50a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.448357] env[63175]: DEBUG nova.compute.provider_tree [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.453437] env[63175]: DEBUG nova.network.neutron [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.547646] env[63175]: DEBUG nova.network.neutron [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.669125] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24604} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.669654] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 544.670862] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25dd105-48eb-4ee4-89fc-20aa99422d41 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.696389] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 544.696694] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6189a93a-ccd0-42e1-b644-d4be55fa4bae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.716757] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 544.716757] env[63175]: value = "task-1247862" [ 544.716757] env[63175]: _type = "Task" [ 544.716757] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.724908] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.856303] env[63175]: DEBUG nova.network.neutron [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.927718] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0e9d9db1-1cc5-487c-b328-290991822d98 tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.877s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.958263] env[63175]: DEBUG nova.scheduler.client.report [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 545.050643] env[63175]: DEBUG oslo_concurrency.lockutils [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] Releasing lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.051026] env[63175]: DEBUG nova.compute.manager [req-b694e388-114f-4ced-8d86-31115140402c req-ea7bd961-7c11-4ce9-b0fd-c2fdf85d19c7 service nova] [instance: 1561f155-ff79-4790-a164-43f08230e229] Received event network-vif-deleted-231db14b-d441-4dc0-aaa3-3534e92fb7e4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 545.056545] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquiring lock "7d227eb2-b270-4304-91ee-7ae20c33f633" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.056836] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "7d227eb2-b270-4304-91ee-7ae20c33f633" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.233463] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247862, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.283835] env[63175]: ERROR nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 545.283835] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.283835] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.283835] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.283835] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.283835] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.283835] env[63175]: ERROR nova.compute.manager raise self.value [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.283835] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.283835] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.283835] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.284368] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.284368] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.284368] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 545.284368] env[63175]: ERROR nova.compute.manager [ 545.284368] env[63175]: Traceback (most recent call last): [ 545.284368] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.284368] env[63175]: listener.cb(fileno) [ 545.284368] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.284368] env[63175]: result = function(*args, **kwargs) [ 545.284368] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.284368] env[63175]: return func(*args, **kwargs) [ 545.284368] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.284368] env[63175]: raise e [ 545.284368] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.284368] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 545.284368] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.284368] env[63175]: created_port_ids = self._update_ports_for_instance( [ 545.284368] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.284368] env[63175]: with excutils.save_and_reraise_exception(): [ 545.284368] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.284368] env[63175]: self.force_reraise() [ 545.284368] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.284368] env[63175]: raise self.value [ 545.284368] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.284368] env[63175]: updated_port = self._update_port( [ 545.284368] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.284368] env[63175]: _ensure_no_port_binding_failure(port) [ 545.284368] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.284368] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.285718] env[63175]: nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 545.285718] env[63175]: Removing descriptor: 15 [ 545.285718] env[63175]: ERROR nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Traceback (most recent call last): [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] yield resources [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.driver.spawn(context, instance, image_meta, [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.285718] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] vm_ref = self.build_virtual_machine(instance, [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] for vif in network_info: [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self._sync_wrapper(fn, *args, **kwargs) [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.wait() [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self[:] = self._gt.wait() [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self._exit_event.wait() [ 545.286226] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] result = hub.switch() [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self.greenlet.switch() [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] result = function(*args, **kwargs) [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return func(*args, **kwargs) [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise e [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] nwinfo = self.network_api.allocate_for_instance( [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.286596] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] created_port_ids = self._update_ports_for_instance( [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] with excutils.save_and_reraise_exception(): [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.force_reraise() [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise self.value [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] updated_port = self._update_port( [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] _ensure_no_port_binding_failure(port) [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.286910] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise exception.PortBindingFailed(port_id=port['id']) [ 545.287253] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 545.287253] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] [ 545.287253] env[63175]: INFO nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Terminating instance [ 545.362700] env[63175]: INFO nova.compute.manager [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] [instance: 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52] Took 1.04 seconds to deallocate network for instance. [ 545.433388] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 545.463945] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.464356] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 545.467924] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.787s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.470047] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.470047] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 545.470047] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.713s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.477051] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f842f7ae-cee3-4476-b325-3168904eae16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.491018] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0647dbec-9ea4-404e-aa2c-c6f5321e3893 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.509700] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83f5846-1daf-4013-9596-d713bd802140 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.516857] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927ca8ac-5b7e-4ab2-9d42-edad632dac90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.549569] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181547MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 545.549678] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.731402] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247862, 'name': ReconfigVM_Task, 'duration_secs': 0.914028} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.731585] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 545.732218] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09d66756-a923-4946-a296-360ea28b305a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.740707] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 545.740707] env[63175]: value = "task-1247863" [ 545.740707] env[63175]: _type = "Task" [ 545.740707] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.751461] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247863, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.793560] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquiring lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.793560] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquired lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.793560] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.962268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.975404] env[63175]: DEBUG nova.compute.utils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 545.979329] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 545.980084] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.048023] env[63175]: DEBUG nova.policy [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b121c6d8b396422e867117e1a3743033', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80e59890e8c7492daaac854ba69a7c4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.261428] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247863, 'name': Rename_Task, 'duration_secs': 0.214626} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.261428] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 546.261706] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96aa20d6-2029-43b2-ad32-60b0eb130ad3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.265287] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1ddd4c-7010-4e3b-9e6e-522b10ece799 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.277109] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 546.277109] env[63175]: value = "task-1247864" [ 546.277109] env[63175]: _type = "Task" [ 546.277109] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.280013] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55caa207-772e-44e3-96b7-178d650eaa70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.291898] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247864, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.322938] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15cef4e-0bb9-4396-b1ea-b2e3613b2dcf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.332309] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe98931a-47f7-4d7d-a159-2f96a5ac2b14 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.345552] env[63175]: DEBUG nova.compute.provider_tree [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.351213] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquiring lock "3c50a11c-cab7-461f-9550-575c4f51a4e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.351556] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "3c50a11c-cab7-461f-9550-575c4f51a4e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.356157] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.401716] env[63175]: INFO nova.scheduler.client.report [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Deleted allocations for instance 0e5d3fa1-19ea-4458-974f-4cb76e7d0b52 [ 546.486498] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 546.515095] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.546107] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Successfully created port: 091accda-b313-467a-8f66-b0db11838464 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.796718] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247864, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.856496] env[63175]: DEBUG nova.scheduler.client.report [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 546.897372] env[63175]: DEBUG nova.compute.manager [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Received event network-vif-deleted-d59a71f1-94b2-4955-bd65-ae6e532707ae {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 546.897372] env[63175]: DEBUG nova.compute.manager [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Received event network-changed-c79ee925-f64d-4b19-b0c7-d4b8214c72f9 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 546.897372] env[63175]: DEBUG nova.compute.manager [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Refreshing instance network info cache due to event network-changed-c79ee925-f64d-4b19-b0c7-d4b8214c72f9. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 546.897372] env[63175]: DEBUG oslo_concurrency.lockutils [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] Acquiring lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.915934] env[63175]: DEBUG oslo_concurrency.lockutils [None req-719dc8a5-4d4b-4078-9c09-17ecddaffefd tempest-ServerDiagnosticsTest-1179899089 tempest-ServerDiagnosticsTest-1179899089-project-member] Lock "0e5d3fa1-19ea-4458-974f-4cb76e7d0b52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.526s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.017147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Releasing lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.019400] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.019603] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.019928] env[63175]: DEBUG oslo_concurrency.lockutils [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] Acquired lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.020485] env[63175]: DEBUG nova.network.neutron [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Refreshing network info cache for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 547.021233] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5c1c5e2-3637-4567-b6e3-34d76efb6d28 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.035776] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce8ed8c-20c3-4004-b81f-8e4336fd2d33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.068986] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9595c38c-3699-4b89-bf37-b2e435886878 could not be found. [ 547.069286] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.069517] env[63175]: INFO nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.069767] env[63175]: DEBUG oslo.service.loopingcall [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.070014] env[63175]: DEBUG nova.compute.manager [-] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.071270] env[63175]: DEBUG nova.network.neutron [-] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.088122] env[63175]: DEBUG nova.network.neutron [-] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.292899] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247864, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.359962] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.363610] env[63175]: ERROR nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Traceback (most recent call last): [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.driver.spawn(context, instance, image_meta, [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] vm_ref = self.build_virtual_machine(instance, [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.363610] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] for vif in network_info: [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self._sync_wrapper(fn, *args, **kwargs) [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.wait() [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self[:] = self._gt.wait() [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self._exit_event.wait() [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] result = hub.switch() [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.363978] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return self.greenlet.switch() [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] result = function(*args, **kwargs) [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] return func(*args, **kwargs) [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise e [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] nwinfo = self.network_api.allocate_for_instance( [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] created_port_ids = self._update_ports_for_instance( [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] with excutils.save_and_reraise_exception(): [ 547.364309] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] self.force_reraise() [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise self.value [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] updated_port = self._update_port( [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] _ensure_no_port_binding_failure(port) [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] raise exception.PortBindingFailed(port_id=port['id']) [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] nova.exception.PortBindingFailed: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. [ 547.364681] env[63175]: ERROR nova.compute.manager [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] [ 547.365041] env[63175]: DEBUG nova.compute.utils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.365041] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.891s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.366902] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Build of instance a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf was re-scheduled: Binding failed for port edfa6f97-302b-41e5-8c84-a3add567630a, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 547.367386] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 547.367623] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquiring lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.367769] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Acquired lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.367927] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.419717] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 547.500188] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 547.528156] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquiring lock "652f4c1b-69b4-415d-ab14-bbf10043471d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.531063] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "652f4c1b-69b4-415d-ab14-bbf10043471d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.541063] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 547.541319] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 547.541475] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.541641] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 547.541755] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.541928] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 547.542187] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 547.542467] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 547.542649] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 547.542706] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 547.542877] env[63175]: DEBUG nova.virt.hardware [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 547.545575] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d589370-2a1d-432f-8082-a2db94265770 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.553518] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70692f3c-da7c-42d7-ae0a-97caff9ba8da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.571515] env[63175]: DEBUG nova.network.neutron [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.594068] env[63175]: DEBUG nova.network.neutron [-] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.755357] env[63175]: DEBUG nova.network.neutron [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.796206] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247864, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.889181] env[63175]: ERROR nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 547.889181] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.889181] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.889181] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.889181] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.889181] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.889181] env[63175]: ERROR nova.compute.manager raise self.value [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.889181] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.889181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.889181] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.889654] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.889654] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.889654] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 547.889654] env[63175]: ERROR nova.compute.manager [ 547.889654] env[63175]: Traceback (most recent call last): [ 547.889654] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.889654] env[63175]: listener.cb(fileno) [ 547.889654] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.889654] env[63175]: result = function(*args, **kwargs) [ 547.889654] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.889654] env[63175]: return func(*args, **kwargs) [ 547.889654] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.889654] env[63175]: raise e [ 547.889654] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.889654] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 547.889654] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.889654] env[63175]: created_port_ids = self._update_ports_for_instance( [ 547.889654] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.889654] env[63175]: with excutils.save_and_reraise_exception(): [ 547.889654] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.889654] env[63175]: self.force_reraise() [ 547.889654] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.889654] env[63175]: raise self.value [ 547.889654] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.889654] env[63175]: updated_port = self._update_port( [ 547.889654] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.889654] env[63175]: _ensure_no_port_binding_failure(port) [ 547.889654] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.889654] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.890471] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 547.890471] env[63175]: Removing descriptor: 15 [ 547.890471] env[63175]: ERROR nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Traceback (most recent call last): [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] yield resources [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.driver.spawn(context, instance, image_meta, [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.890471] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] vm_ref = self.build_virtual_machine(instance, [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] for vif in network_info: [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self._sync_wrapper(fn, *args, **kwargs) [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.wait() [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self[:] = self._gt.wait() [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self._exit_event.wait() [ 547.890836] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] result = hub.switch() [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self.greenlet.switch() [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] result = function(*args, **kwargs) [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return func(*args, **kwargs) [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise e [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] nwinfo = self.network_api.allocate_for_instance( [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.891227] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] created_port_ids = self._update_ports_for_instance( [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] with excutils.save_and_reraise_exception(): [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.force_reraise() [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise self.value [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] updated_port = self._update_port( [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] _ensure_no_port_binding_failure(port) [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.891625] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise exception.PortBindingFailed(port_id=port['id']) [ 547.891954] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 547.891954] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] [ 547.891954] env[63175]: INFO nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Terminating instance [ 547.902382] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.950783] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.098901] env[63175]: INFO nova.compute.manager [-] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Took 1.03 seconds to deallocate network for instance. [ 548.101853] env[63175]: DEBUG nova.compute.claims [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.102048] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.117876] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.125139] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ee4218-44f8-4b37-a840-14458541cf12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.132677] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b59860-edeb-4312-b9d0-c42e486b20fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.165815] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecdb76f-58e0-4058-b3ee-3045e09fb632 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.173352] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd978d9-75fe-4d31-912f-f165d8870a6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.187021] env[63175]: DEBUG nova.compute.provider_tree [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.196381] env[63175]: DEBUG nova.compute.manager [None req-4541ac3c-80ae-4c83-8677-4318fb4067a4 tempest-ServerDiagnosticsV248Test-1715394790 tempest-ServerDiagnosticsV248Test-1715394790-project-admin] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 548.198324] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888e0256-8d79-4585-a6e7-6ecacc0419fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.205282] env[63175]: INFO nova.compute.manager [None req-4541ac3c-80ae-4c83-8677-4318fb4067a4 tempest-ServerDiagnosticsV248Test-1715394790 tempest-ServerDiagnosticsV248Test-1715394790-project-admin] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Retrieving diagnostics [ 548.206064] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516fc27a-f24e-4c5f-8575-5c7e654ab658 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.240383] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.240608] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.256115] env[63175]: DEBUG oslo_concurrency.lockutils [req-504f33b4-a524-4dd7-9d8f-bc2c12e7e38d req-a0b975ef-30f6-4fdf-8eab-7e7e5329d2fd service nova] Releasing lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.292613] env[63175]: DEBUG oslo_vmware.api [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247864, 'name': PowerOnVM_Task, 'duration_secs': 1.651804} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.292863] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 548.293071] env[63175]: DEBUG nova.compute.manager [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 548.293799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705bba81-98f4-41a4-bab5-f423f939a287 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.397932] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquiring lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.398140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquired lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.398367] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.620836] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Releasing lock "refresh_cache-a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.621238] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 548.621444] env[63175]: DEBUG nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 548.621768] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.646400] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.690226] env[63175]: DEBUG nova.scheduler.client.report [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 548.808544] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.924767] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.077658] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.150478] env[63175]: DEBUG nova.network.neutron [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.194602] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.832s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.195492] env[63175]: ERROR nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Traceback (most recent call last): [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.driver.spawn(context, instance, image_meta, [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] vm_ref = self.build_virtual_machine(instance, [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.195492] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] for vif in network_info: [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self._sync_wrapper(fn, *args, **kwargs) [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.wait() [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self[:] = self._gt.wait() [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self._exit_event.wait() [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] result = hub.switch() [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.195866] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return self.greenlet.switch() [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] result = function(*args, **kwargs) [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] return func(*args, **kwargs) [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise e [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] nwinfo = self.network_api.allocate_for_instance( [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] created_port_ids = self._update_ports_for_instance( [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] with excutils.save_and_reraise_exception(): [ 549.196233] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] self.force_reraise() [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise self.value [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] updated_port = self._update_port( [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] _ensure_no_port_binding_failure(port) [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] raise exception.PortBindingFailed(port_id=port['id']) [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] nova.exception.PortBindingFailed: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. [ 549.196595] env[63175]: ERROR nova.compute.manager [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] [ 549.198529] env[63175]: DEBUG nova.compute.utils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.199874] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.580s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.200793] env[63175]: INFO nova.compute.claims [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.204230] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Build of instance 18acf9cd-0661-48c0-8fd2-b80edba4dfa6 was re-scheduled: Binding failed for port 0d58a188-6ddc-4364-9486-421c86294015, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 549.207621] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 549.207621] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquiring lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.207621] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Acquired lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.207621] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.584272] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Releasing lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.584272] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 549.584272] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 549.584272] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9849839-3c2b-42bd-9c56-5d286e14c36c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.596993] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cb4cad-077f-4fe5-968a-4daa693ee784 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.623083] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 596f899b-18f2-42f0-968e-6e9159f1d4d6 could not be found. [ 549.623083] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.623083] env[63175]: INFO nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 549.623681] env[63175]: DEBUG oslo.service.loopingcall [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.623927] env[63175]: DEBUG nova.compute.manager [-] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 549.624082] env[63175]: DEBUG nova.network.neutron [-] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.650434] env[63175]: DEBUG nova.network.neutron [-] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.654883] env[63175]: INFO nova.compute.manager [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] [instance: a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf] Took 1.03 seconds to deallocate network for instance. [ 549.730971] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.742393] env[63175]: DEBUG nova.compute.manager [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Received event network-vif-deleted-c79ee925-f64d-4b19-b0c7-d4b8214c72f9 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 549.742742] env[63175]: DEBUG nova.compute.manager [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Received event network-changed-091accda-b313-467a-8f66-b0db11838464 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 549.742978] env[63175]: DEBUG nova.compute.manager [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Refreshing instance network info cache due to event network-changed-091accda-b313-467a-8f66-b0db11838464. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 549.743144] env[63175]: DEBUG oslo_concurrency.lockutils [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] Acquiring lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.743261] env[63175]: DEBUG oslo_concurrency.lockutils [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] Acquired lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.743402] env[63175]: DEBUG nova.network.neutron [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Refreshing network info cache for port 091accda-b313-467a-8f66-b0db11838464 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 549.824456] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.153104] env[63175]: DEBUG nova.network.neutron [-] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.270788] env[63175]: DEBUG nova.network.neutron [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.328581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Releasing lock "refresh_cache-18acf9cd-0661-48c0-8fd2-b80edba4dfa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.329126] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 550.329354] env[63175]: DEBUG nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 550.329565] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.352721] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.498472] env[63175]: DEBUG nova.network.neutron [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.529755] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fb9ad7-334a-4120-91d4-b1fafc1a0c02 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.540686] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f2238c-f320-4f87-9080-30f68375cdf2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.577926] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39c6e5d-7517-4a88-ade7-5b66e35463c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.585902] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80142c0b-60f1-4575-99b2-8285555da63c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.602506] env[63175]: DEBUG nova.compute.provider_tree [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.656576] env[63175]: INFO nova.compute.manager [-] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Took 1.03 seconds to deallocate network for instance. [ 550.664614] env[63175]: DEBUG nova.compute.claims [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 550.664852] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.700658] env[63175]: INFO nova.scheduler.client.report [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Deleted allocations for instance a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf [ 550.861029] env[63175]: DEBUG nova.network.neutron [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.001756] env[63175]: DEBUG oslo_concurrency.lockutils [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] Releasing lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.002028] env[63175]: DEBUG nova.compute.manager [req-77c05cbc-e350-41c8-a4ce-4fe31ba0203b req-cc2d13fe-8b9e-4d91-98c2-56abac12c432 service nova] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Received event network-vif-deleted-091accda-b313-467a-8f66-b0db11838464 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 551.106741] env[63175]: DEBUG nova.scheduler.client.report [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 551.219701] env[63175]: INFO nova.compute.manager [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Rebuilding instance [ 551.222060] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e711f466-5f85-4045-a563-d177b5222143 tempest-ImagesNegativeTestJSON-1067895377 tempest-ImagesNegativeTestJSON-1067895377-project-member] Lock "a7919094-1f9a-46e1-a0fc-6b7ae07fb6cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.341s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.274713] env[63175]: DEBUG nova.compute.manager [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 551.275614] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889e6fc9-af9c-4a6d-b152-41b6f5561239 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.366095] env[63175]: INFO nova.compute.manager [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] [instance: 18acf9cd-0661-48c0-8fd2-b80edba4dfa6] Took 1.04 seconds to deallocate network for instance. [ 551.615471] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.615471] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 551.622685] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.880s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.725032] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 552.091498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "4883dd7f-a13e-4efc-90ff-ca42909a7f0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.091837] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "4883dd7f-a13e-4efc-90ff-ca42909a7f0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.131592] env[63175]: DEBUG nova.compute.utils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.136417] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 552.136584] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 552.247704] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.292719] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 552.293365] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eea860be-4286-4554-a291-fc8dba4f64ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.301096] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 552.301096] env[63175]: value = "task-1247865" [ 552.301096] env[63175]: _type = "Task" [ 552.301096] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.309338] env[63175]: DEBUG nova.policy [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e0bf05d334444e4ca1d35ae10c972ec7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8e558fa473141a68a95719bec89b2f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.315476] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.395823] env[63175]: INFO nova.scheduler.client.report [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Deleted allocations for instance 18acf9cd-0661-48c0-8fd2-b80edba4dfa6 [ 552.436142] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f40d02-5168-40c7-aeb3-5321efe9078d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.446849] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3311490c-e040-4d3b-9f0c-21a4ba099826 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.480936] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d786f4-5b09-4fe5-a5a3-8e30467f6dfe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.488752] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5409906b-155a-4c67-b438-35644fc908db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.502505] env[63175]: DEBUG nova.compute.provider_tree [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.642164] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 552.814638] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247865, 'name': PowerOffVM_Task, 'duration_secs': 0.124264} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.815039] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 552.816445] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.817360] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c32d3b-aa8d-4b94-9bae-d68dfa362af7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.824765] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 552.824877] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f740c7a5-08ee-4822-9be9-dc0a65508153 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.848811] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 552.848911] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 552.849158] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Deleting the datastore file [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 552.849445] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-868dc051-538d-4878-b733-3892f280b2f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.858595] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 552.858595] env[63175]: value = "task-1247867" [ 552.858595] env[63175]: _type = "Task" [ 552.858595] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.868635] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.910652] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d7fe61f-e358-4c17-9992-28b83c9cc0ed tempest-ServersAdminNegativeTestJSON-925900215 tempest-ServersAdminNegativeTestJSON-925900215-project-member] Lock "18acf9cd-0661-48c0-8fd2-b80edba4dfa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.377s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.008028] env[63175]: DEBUG nova.scheduler.client.report [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 553.370612] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099628} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.370930] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 553.371946] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 553.371946] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.413366] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 553.497242] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "8fa60fe5-a397-4b57-bb18-f41c0029743b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.497484] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "8fa60fe5-a397-4b57-bb18-f41c0029743b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.498877] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Successfully created port: f88926f0-7636-4e1c-b295-ad32eb41e10f {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.512845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.513479] env[63175]: ERROR nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Traceback (most recent call last): [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.driver.spawn(context, instance, image_meta, [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] vm_ref = self.build_virtual_machine(instance, [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.513479] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] for vif in network_info: [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self._sync_wrapper(fn, *args, **kwargs) [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.wait() [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self[:] = self._gt.wait() [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self._exit_event.wait() [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] result = hub.switch() [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.513825] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return self.greenlet.switch() [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] result = function(*args, **kwargs) [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] return func(*args, **kwargs) [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise e [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] nwinfo = self.network_api.allocate_for_instance( [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] created_port_ids = self._update_ports_for_instance( [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] with excutils.save_and_reraise_exception(): [ 553.514209] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] self.force_reraise() [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise self.value [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] updated_port = self._update_port( [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] _ensure_no_port_binding_failure(port) [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] raise exception.PortBindingFailed(port_id=port['id']) [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] nova.exception.PortBindingFailed: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. [ 553.514593] env[63175]: ERROR nova.compute.manager [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] [ 553.514916] env[63175]: DEBUG nova.compute.utils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.515513] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.421s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.518838] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Build of instance 380a6278-22d4-485b-b9cb-ccb1ac07d4a9 was re-scheduled: Binding failed for port 4d86bf78-20e6-4fd7-bdcc-6fe89704e13d, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 553.520605] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 553.520605] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.520605] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquired lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.520605] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.657974] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 553.694248] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.694533] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.694820] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.694996] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.695089] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.695246] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.695482] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.695885] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.695885] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.695951] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.696083] env[63175]: DEBUG nova.virt.hardware [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.697129] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d880369e-a3b3-4f77-9327-49149aa32b07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.707407] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca5aa5b-7b18-477d-bed6-8e1a96eef5d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.938396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.083847] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.303063] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.328028] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e4a7ca-6056-40de-b8a2-0b225cdff740 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.339236] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309612d1-7da7-4821-8e28-8557d8a87638 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.376170] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4716b18a-e61e-414d-8571-2c2f16ff054b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.388632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f02a7c-b436-4298-9e00-36e209f69984 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.407534] env[63175]: DEBUG nova.compute.provider_tree [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.423030] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.423186] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.423280] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.423513] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.423599] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.423907] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.424351] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.424571] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.425639] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.425639] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.425639] env[63175]: DEBUG nova.virt.hardware [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.425940] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f88f70-bc1b-4944-a856-ab6e8e8ac932 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.435608] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052c6192-b468-49d5-85a6-baca2402bbbd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.451807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 554.457396] env[63175]: DEBUG oslo.service.loopingcall [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.461297] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 554.461297] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5651964-a2a3-4022-a7e6-2a6554414758 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.479393] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 554.479393] env[63175]: value = "task-1247868" [ 554.479393] env[63175]: _type = "Task" [ 554.479393] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.487386] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247868, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.805176] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Releasing lock "refresh_cache-380a6278-22d4-485b-b9cb-ccb1ac07d4a9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.805455] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 554.805623] env[63175]: DEBUG nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 554.805791] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.853799] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.912420] env[63175]: DEBUG nova.scheduler.client.report [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 554.993713] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247868, 'name': CreateVM_Task, 'duration_secs': 0.290077} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.993713] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 554.993961] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.994131] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.994445] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 554.994689] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b50ee73-b048-4f04-93c6-0f73a3eeda76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.000140] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 555.000140] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fcf2ff-0779-7efa-186d-f7ca01202241" [ 555.000140] env[63175]: _type = "Task" [ 555.000140] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.008599] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fcf2ff-0779-7efa-186d-f7ca01202241, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.235637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquiring lock "dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.235932] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.358311] env[63175]: DEBUG nova.network.neutron [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.418256] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.418883] env[63175]: ERROR nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] Traceback (most recent call last): [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.driver.spawn(context, instance, image_meta, [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] vm_ref = self.build_virtual_machine(instance, [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.418883] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] for vif in network_info: [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self._sync_wrapper(fn, *args, **kwargs) [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.wait() [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self[:] = self._gt.wait() [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self._exit_event.wait() [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] result = hub.switch() [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.419371] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return self.greenlet.switch() [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] result = function(*args, **kwargs) [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] return func(*args, **kwargs) [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise e [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] nwinfo = self.network_api.allocate_for_instance( [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] created_port_ids = self._update_ports_for_instance( [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] with excutils.save_and_reraise_exception(): [ 555.419784] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] self.force_reraise() [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise self.value [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] updated_port = self._update_port( [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] _ensure_no_port_binding_failure(port) [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] raise exception.PortBindingFailed(port_id=port['id']) [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] nova.exception.PortBindingFailed: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. [ 555.420120] env[63175]: ERROR nova.compute.manager [instance: 1561f155-ff79-4790-a164-43f08230e229] [ 555.420391] env[63175]: DEBUG nova.compute.utils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.422814] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.397s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.425783] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Build of instance 1561f155-ff79-4790-a164-43f08230e229 was re-scheduled: Binding failed for port 231db14b-d441-4dc0-aaa3-3534e92fb7e4, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 555.426337] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 555.426417] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquiring lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.426822] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Acquired lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.426822] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.512488] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fcf2ff-0779-7efa-186d-f7ca01202241, 'name': SearchDatastore_Task, 'duration_secs': 0.008521} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.513435] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.513435] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 555.513435] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.513435] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.513590] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.513788] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7579b63-d000-4c65-9089-ddbb2951d68c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.521892] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.522094] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 555.522910] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbbfb28f-182e-4d27-9a96-8fa91d494624 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.528428] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 555.528428] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5238b5a8-ee3f-a402-8eb3-dff6a27c5f1e" [ 555.528428] env[63175]: _type = "Task" [ 555.528428] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.536281] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5238b5a8-ee3f-a402-8eb3-dff6a27c5f1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.860757] env[63175]: INFO nova.compute.manager [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 380a6278-22d4-485b-b9cb-ccb1ac07d4a9] Took 1.05 seconds to deallocate network for instance. [ 556.013760] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.036490] env[63175]: DEBUG nova.compute.manager [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Received event network-changed-f88926f0-7636-4e1c-b295-ad32eb41e10f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 556.036689] env[63175]: DEBUG nova.compute.manager [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Refreshing instance network info cache due to event network-changed-f88926f0-7636-4e1c-b295-ad32eb41e10f. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 556.037693] env[63175]: DEBUG oslo_concurrency.lockutils [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] Acquiring lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.037693] env[63175]: DEBUG oslo_concurrency.lockutils [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] Acquired lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.037693] env[63175]: DEBUG nova.network.neutron [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Refreshing network info cache for port f88926f0-7636-4e1c-b295-ad32eb41e10f {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 556.048407] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5238b5a8-ee3f-a402-8eb3-dff6a27c5f1e, 'name': SearchDatastore_Task, 'duration_secs': 0.008512} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.056379] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c939bba7-0a5e-4b32-9d6e-d49258d9d135 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.063398] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 556.063398] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c50df1-e473-6486-c643-62b9cc994d30" [ 556.063398] env[63175]: _type = "Task" [ 556.063398] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.073073] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c50df1-e473-6486-c643-62b9cc994d30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.241434] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4828c1c4-f9b5-466b-adb5-9e87a46946bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.246139] env[63175]: ERROR nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 556.246139] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.246139] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.246139] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.246139] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.246139] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.246139] env[63175]: ERROR nova.compute.manager raise self.value [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.246139] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.246139] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.246139] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.246621] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.246621] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.246621] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 556.246621] env[63175]: ERROR nova.compute.manager [ 556.246621] env[63175]: Traceback (most recent call last): [ 556.246621] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.246621] env[63175]: listener.cb(fileno) [ 556.246621] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.246621] env[63175]: result = function(*args, **kwargs) [ 556.246621] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.246621] env[63175]: return func(*args, **kwargs) [ 556.246621] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.246621] env[63175]: raise e [ 556.246621] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.246621] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 556.246621] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.246621] env[63175]: created_port_ids = self._update_ports_for_instance( [ 556.246621] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.246621] env[63175]: with excutils.save_and_reraise_exception(): [ 556.246621] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.246621] env[63175]: self.force_reraise() [ 556.246621] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.246621] env[63175]: raise self.value [ 556.246621] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.246621] env[63175]: updated_port = self._update_port( [ 556.246621] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.246621] env[63175]: _ensure_no_port_binding_failure(port) [ 556.246621] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.246621] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.247522] env[63175]: nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 556.247522] env[63175]: Removing descriptor: 17 [ 556.251276] env[63175]: ERROR nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Traceback (most recent call last): [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] yield resources [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.driver.spawn(context, instance, image_meta, [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] vm_ref = self.build_virtual_machine(instance, [ 556.251276] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] for vif in network_info: [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self._sync_wrapper(fn, *args, **kwargs) [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.wait() [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self[:] = self._gt.wait() [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self._exit_event.wait() [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.252052] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] result = hub.switch() [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self.greenlet.switch() [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] result = function(*args, **kwargs) [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return func(*args, **kwargs) [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise e [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] nwinfo = self.network_api.allocate_for_instance( [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] created_port_ids = self._update_ports_for_instance( [ 556.252605] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] with excutils.save_and_reraise_exception(): [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.force_reraise() [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise self.value [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] updated_port = self._update_port( [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] _ensure_no_port_binding_failure(port) [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise exception.PortBindingFailed(port_id=port['id']) [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 556.253745] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] [ 556.255646] env[63175]: INFO nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Terminating instance [ 556.255646] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.269354] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af7d1bc-7550-4c00-a59b-59c3bd8796a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.305486] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddc828c-d2e8-4363-b724-df476ddedaae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.314470] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d677a6c-a654-4da1-a058-8647aeaba041 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.329914] env[63175]: DEBUG nova.compute.provider_tree [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.571238] env[63175]: DEBUG nova.network.neutron [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.579528] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c50df1-e473-6486-c643-62b9cc994d30, 'name': SearchDatastore_Task, 'duration_secs': 0.009212} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.580474] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.580474] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 556.580474] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd7ce100-7c7b-482a-90e1-394da9f75580 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.591547] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 556.591547] env[63175]: value = "task-1247869" [ 556.591547] env[63175]: _type = "Task" [ 556.591547] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.601543] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.722610] env[63175]: DEBUG nova.network.neutron [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.761318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Releasing lock "refresh_cache-1561f155-ff79-4790-a164-43f08230e229" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.761544] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 556.761702] env[63175]: DEBUG nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 556.761903] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.766335] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.827806] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.836023] env[63175]: DEBUG nova.scheduler.client.report [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 556.921426] env[63175]: INFO nova.scheduler.client.report [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Deleted allocations for instance 380a6278-22d4-485b-b9cb-ccb1ac07d4a9 [ 557.105660] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247869, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.228726] env[63175]: DEBUG oslo_concurrency.lockutils [req-2f12e7ab-d2f8-4f11-b146-75dddaf0ae2e req-cd352a68-4dff-4623-b645-4ef768d48df7 service nova] Releasing lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.231823] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquired lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.231823] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.330727] env[63175]: DEBUG nova.network.neutron [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.344022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.344022] env[63175]: ERROR nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Traceback (most recent call last): [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.driver.spawn(context, instance, image_meta, [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.344022] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] vm_ref = self.build_virtual_machine(instance, [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] for vif in network_info: [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self._sync_wrapper(fn, *args, **kwargs) [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.wait() [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self[:] = self._gt.wait() [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self._exit_event.wait() [ 557.344619] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] result = hub.switch() [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return self.greenlet.switch() [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] result = function(*args, **kwargs) [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] return func(*args, **kwargs) [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise e [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] nwinfo = self.network_api.allocate_for_instance( [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.344968] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] created_port_ids = self._update_ports_for_instance( [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] with excutils.save_and_reraise_exception(): [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] self.force_reraise() [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise self.value [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] updated_port = self._update_port( [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] _ensure_no_port_binding_failure(port) [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.345352] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] raise exception.PortBindingFailed(port_id=port['id']) [ 557.345678] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] nova.exception.PortBindingFailed: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. [ 557.345678] env[63175]: ERROR nova.compute.manager [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] [ 557.345678] env[63175]: DEBUG nova.compute.utils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.345678] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.795s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.346972] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Build of instance 0e0f0535-d539-4572-8752-e3e6a4b79b8c was re-scheduled: Binding failed for port d59a71f1-94b2-4955-bd65-ae6e532707ae, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 557.346972] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 557.347161] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquiring lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.347161] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Acquired lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.348642] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.435283] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e74b1895-aac7-4aaa-b6df-f9c58f31a9c4 tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "380a6278-22d4-485b-b9cb-ccb1ac07d4a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.237s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.604508] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695558} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.604508] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 557.604508] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 557.604508] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af999c79-190a-49b4-bd1d-4d1e85c7d172 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.613060] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 557.613060] env[63175]: value = "task-1247870" [ 557.613060] env[63175]: _type = "Task" [ 557.613060] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.633221] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.838878] env[63175]: INFO nova.compute.manager [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] [instance: 1561f155-ff79-4790-a164-43f08230e229] Took 1.08 seconds to deallocate network for instance. [ 557.925773] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.937136] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 557.971681] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.058438] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.129918] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106781} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.130382] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 558.132628] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb477237-988e-4377-9d5b-36660b354cd6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.153454] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 558.153780] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-291d7dba-36d9-46e8-b9ee-133b52ab66a0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.175289] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 558.175289] env[63175]: value = "task-1247871" [ 558.175289] env[63175]: _type = "Task" [ 558.175289] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.183990] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.286857] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.293742] env[63175]: DEBUG nova.compute.manager [req-f55eb57c-07e7-4563-964b-9e54a09e831a req-38d656c4-5306-4b54-9da3-3296468eaaaa service nova] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Received event network-vif-deleted-f88926f0-7636-4e1c-b295-ad32eb41e10f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 558.398167] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 24ee74d8-33d2-4efa-97de-a642c89a3461 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.400030] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 56b3ba41-0406-4b69-89af-b00e87feaa89 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.470483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.561300] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Releasing lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.561726] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 558.561911] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.562234] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7835573-5c78-4545-8a4a-8a0446f78436 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.576350] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b692d02-83fe-4317-83c3-8b9b53b03581 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.616216] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e96543c8-dc4e-45b3-8ece-297ee6dbcedb could not be found. [ 558.616710] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.617636] env[63175]: INFO nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Took 0.06 seconds to destroy the instance on the hypervisor. [ 558.617935] env[63175]: DEBUG oslo.service.loopingcall [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.618442] env[63175]: DEBUG nova.compute.manager [-] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 558.618442] env[63175]: DEBUG nova.network.neutron [-] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.686158] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247871, 'name': ReconfigVM_Task, 'duration_secs': 0.299796} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.686461] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461/24ee74d8-33d2-4efa-97de-a642c89a3461.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 558.687148] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-644af66e-db15-4ad6-92b0-5b73d3f13614 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.694789] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 558.694789] env[63175]: value = "task-1247872" [ 558.694789] env[63175]: _type = "Task" [ 558.694789] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.707573] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247872, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.794921] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Releasing lock "refresh_cache-0e0f0535-d539-4572-8752-e3e6a4b79b8c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.795180] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 558.795377] env[63175]: DEBUG nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 558.795588] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.833999] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.851940] env[63175]: DEBUG nova.network.neutron [-] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.895091] env[63175]: INFO nova.scheduler.client.report [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Deleted allocations for instance 1561f155-ff79-4790-a164-43f08230e229 [ 558.904846] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 1561f155-ff79-4790-a164-43f08230e229 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.097895] env[63175]: DEBUG nova.compute.manager [None req-0db357f6-b978-4322-bdda-0e9f92a87e87 tempest-ServerDiagnosticsV248Test-1715394790 tempest-ServerDiagnosticsV248Test-1715394790-project-admin] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 559.098999] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76dcd734-c239-45ce-ae3c-ab9da2708ac7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.107351] env[63175]: INFO nova.compute.manager [None req-0db357f6-b978-4322-bdda-0e9f92a87e87 tempest-ServerDiagnosticsV248Test-1715394790 tempest-ServerDiagnosticsV248Test-1715394790-project-admin] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Retrieving diagnostics [ 559.108410] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ab4364-0cd5-4102-a6aa-a4a83efaa930 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.205477] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247872, 'name': Rename_Task, 'duration_secs': 0.160807} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.205477] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 559.205630] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7548e87-5d0a-461d-b3c6-cc5c311fd7c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.212169] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Waiting for the task: (returnval){ [ 559.212169] env[63175]: value = "task-1247873" [ 559.212169] env[63175]: _type = "Task" [ 559.212169] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.221776] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.337277] env[63175]: DEBUG nova.network.neutron [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.359141] env[63175]: DEBUG nova.network.neutron [-] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.406823] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0e0f0535-d539-4572-8752-e3e6a4b79b8c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.407094] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9595c38c-3699-4b89-bf37-b2e435886878 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.407184] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 596f899b-18f2-42f0-968e-6e9159f1d4d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.407706] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e96543c8-dc4e-45b3-8ece-297ee6dbcedb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.411115] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23c47816-9abb-49f1-9e43-cfaf8026d32a tempest-ServerExternalEventsTest-940031504 tempest-ServerExternalEventsTest-940031504-project-member] Lock "1561f155-ff79-4790-a164-43f08230e229" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.526s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.729746] env[63175]: DEBUG oslo_vmware.api [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Task: {'id': task-1247873, 'name': PowerOnVM_Task, 'duration_secs': 0.500655} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.731516] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 559.731516] env[63175]: DEBUG nova.compute.manager [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 559.733086] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c9bdd0-878b-48cf-9605-c3fd1a220d76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.841403] env[63175]: INFO nova.compute.manager [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] [instance: 0e0f0535-d539-4572-8752-e3e6a4b79b8c] Took 1.04 seconds to deallocate network for instance. [ 559.860997] env[63175]: INFO nova.compute.manager [-] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Took 1.24 seconds to deallocate network for instance. [ 559.865121] env[63175]: DEBUG nova.compute.claims [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.865716] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.911921] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance f33d2617-c4ef-4174-a838-ce0558867b9d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.915643] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 560.255856] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.421415] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance edd63563-8fe3-4280-9300-8a11ba40ba68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.440609] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.516801] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "1939e74b-3ab5-428b-86c5-4d6de694c454" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.516801] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "1939e74b-3ab5-428b-86c5-4d6de694c454" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.897824] env[63175]: INFO nova.scheduler.client.report [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Deleted allocations for instance 0e0f0535-d539-4572-8752-e3e6a4b79b8c [ 560.926492] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.040177] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "56b3ba41-0406-4b69-89af-b00e87feaa89" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.040965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.040965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "56b3ba41-0406-4b69-89af-b00e87feaa89-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.041417] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.041504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.044203] env[63175]: INFO nova.compute.manager [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Terminating instance [ 561.409676] env[63175]: DEBUG oslo_concurrency.lockutils [None req-612baeb7-8e5a-4fa4-aafe-e7d1a2ec6c7d tempest-ServersV294TestFqdnHostnames-1476190921 tempest-ServersV294TestFqdnHostnames-1476190921-project-member] Lock "0e0f0535-d539-4572-8752-e3e6a4b79b8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.958s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.432323] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 7d227eb2-b270-4304-91ee-7ae20c33f633 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.550249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "refresh_cache-56b3ba41-0406-4b69-89af-b00e87feaa89" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.550603] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquired lock "refresh_cache-56b3ba41-0406-4b69-89af-b00e87feaa89" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.550824] env[63175]: DEBUG nova.network.neutron [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.735043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "24ee74d8-33d2-4efa-97de-a642c89a3461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.735328] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.735533] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "24ee74d8-33d2-4efa-97de-a642c89a3461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.735802] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.736235] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.738542] env[63175]: INFO nova.compute.manager [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Terminating instance [ 561.792732] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "6cd44d90-9092-483b-ab80-442e8f59435e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.792959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "6cd44d90-9092-483b-ab80-442e8f59435e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.912310] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 561.935363] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 3c50a11c-cab7-461f-9550-575c4f51a4e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.112367] env[63175]: DEBUG nova.network.neutron [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.237472] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "ce379bc8-add6-4008-b4db-c8bdb61a8e97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.237472] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "ce379bc8-add6-4008-b4db-c8bdb61a8e97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.243284] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "refresh_cache-24ee74d8-33d2-4efa-97de-a642c89a3461" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.243394] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquired lock "refresh_cache-24ee74d8-33d2-4efa-97de-a642c89a3461" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.243687] env[63175]: DEBUG nova.network.neutron [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.370463] env[63175]: DEBUG nova.network.neutron [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.441589] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.442099] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 652f4c1b-69b4-415d-ab14-bbf10043471d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.875129] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Releasing lock "refresh_cache-56b3ba41-0406-4b69-89af-b00e87feaa89" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.875573] env[63175]: DEBUG nova.compute.manager [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 562.875764] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.876681] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642176ec-85cd-4459-8b55-1460e3b3dfc4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.890877] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 562.891596] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-160c9921-1c70-4a25-941c-7da3b810409b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.900320] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 562.900320] env[63175]: value = "task-1247874" [ 562.900320] env[63175]: _type = "Task" [ 562.900320] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.909520] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.945463] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.975663] env[63175]: DEBUG nova.network.neutron [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.171886] env[63175]: DEBUG nova.network.neutron [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.413363] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247874, 'name': PowerOffVM_Task, 'duration_secs': 0.129374} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.413884] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 563.414239] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 563.414578] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7f73888-d729-4dd3-9890-81062014dba0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.444786] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 563.445357] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 563.445939] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleting the datastore file [datastore1] 56b3ba41-0406-4b69-89af-b00e87feaa89 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 563.446539] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2204def8-9f36-4833-a568-c3d83d1fb7cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.453914] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 4883dd7f-a13e-4efc-90ff-ca42909a7f0c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.465167] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for the task: (returnval){ [ 563.465167] env[63175]: value = "task-1247876" [ 563.465167] env[63175]: _type = "Task" [ 563.465167] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.484795] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquiring lock "425a1271-cda7-45e4-89cb-cccd6231b623" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.484795] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "425a1271-cda7-45e4-89cb-cccd6231b623" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.491577] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.676400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Releasing lock "refresh_cache-24ee74d8-33d2-4efa-97de-a642c89a3461" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.676621] env[63175]: DEBUG nova.compute.manager [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 563.677027] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.677824] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cabb5bc-ae1d-4be7-a8ad-9c5a4f6ec3a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.688429] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 563.689223] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65e92b36-e48f-41fc-af4b-16340500a030 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.695496] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 563.695496] env[63175]: value = "task-1247877" [ 563.695496] env[63175]: _type = "Task" [ 563.695496] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.704240] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.962890] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 8fa60fe5-a397-4b57-bb18-f41c0029743b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.975217] env[63175]: DEBUG oslo_vmware.api [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Task: {'id': task-1247876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097072} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.975217] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 563.975217] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 563.975217] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.975531] env[63175]: INFO nova.compute.manager [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Took 1.10 seconds to destroy the instance on the hypervisor. [ 563.978817] env[63175]: DEBUG oslo.service.loopingcall [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.978817] env[63175]: DEBUG nova.compute.manager [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 563.978817] env[63175]: DEBUG nova.network.neutron [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.006586] env[63175]: DEBUG nova.network.neutron [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.210788] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247877, 'name': PowerOffVM_Task, 'duration_secs': 0.122505} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.210788] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 564.210788] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 564.210788] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cea0884d-5583-4fb2-b4f3-01f797a42ba5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.240679] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 564.241127] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 564.241399] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleting the datastore file [datastore2] 24ee74d8-33d2-4efa-97de-a642c89a3461 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 564.241770] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-091386ee-9288-4aa8-8abf-8b192ec8fbdd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.248677] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for the task: (returnval){ [ 564.248677] env[63175]: value = "task-1247879" [ 564.248677] env[63175]: _type = "Task" [ 564.248677] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.257035] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.420854] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "5da92241-84f0-4510-b19d-b28cb57079ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.421206] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "5da92241-84f0-4510-b19d-b28cb57079ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.466260] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.467316] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 564.467316] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 564.510239] env[63175]: DEBUG nova.network.neutron [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.719573] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.719815] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.759493] env[63175]: DEBUG oslo_vmware.api [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Task: {'id': task-1247879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090807} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.762347] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.762543] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 564.762721] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.763107] env[63175]: INFO nova.compute.manager [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Took 1.09 seconds to destroy the instance on the hypervisor. [ 564.763185] env[63175]: DEBUG oslo.service.loopingcall [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.763846] env[63175]: DEBUG nova.compute.manager [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 564.763846] env[63175]: DEBUG nova.network.neutron [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.809247] env[63175]: DEBUG nova.network.neutron [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.836018] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c895373-83ee-44c5-a6a0-3e0540ab1a7d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.843628] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456973d0-1f34-436d-b05c-01b9daea86ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.883908] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464e8551-ada4-4f55-a9ec-b45a8565c997 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.893643] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aab23a8-c526-4862-929a-e7296414641a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.909494] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.015565] env[63175]: INFO nova.compute.manager [-] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Took 1.04 seconds to deallocate network for instance. [ 565.311942] env[63175]: DEBUG nova.network.neutron [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.412269] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 565.525980] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.814548] env[63175]: INFO nova.compute.manager [-] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Took 1.05 seconds to deallocate network for instance. [ 565.920396] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 565.920396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.573s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.920396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.956s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.920396] env[63175]: INFO nova.compute.claims [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.326542] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.341400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquiring lock "2cb7696b-f146-4db6-9e83-385fb8af1127" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.341632] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "2cb7696b-f146-4db6-9e83-385fb8af1127" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.303255] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa412817-2934-4762-a5e7-3bba3eeb4a9b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.312032] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9be269-717a-4532-ac1f-d0f178f1a8d3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.353317] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0535fbb-938f-4043-b896-c3272e53042f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.363175] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04394570-a376-4774-82eb-ddd3da7a3fde {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.378358] env[63175]: DEBUG nova.compute.provider_tree [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.884209] env[63175]: DEBUG nova.scheduler.client.report [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 568.389428] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.389428] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 568.395881] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.445s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.402306] env[63175]: INFO nova.compute.claims [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.879300] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquiring lock "d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.879300] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.897569] env[63175]: DEBUG nova.compute.utils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.899247] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 568.899418] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 569.091113] env[63175]: DEBUG nova.policy [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4398621d58394ae5950aec52211263da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d2185912d6b4083a315e01c8e47282c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.403029] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 569.844028] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fd2b2a-4508-4dfe-96e0-58ba7d159ff9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.853026] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d4bcdd-4dd0-4054-9796-000f46bb9433 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.893736] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348519eb-e4e6-4a1a-a0e0-479a7b370647 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.901685] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80dde91d-2a71-40c1-bf53-d9e127b04c40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.920257] env[63175]: DEBUG nova.compute.provider_tree [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.075469] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Successfully created port: d337dd3a-6f73-421e-b660-f27749e51965 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.423293] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 570.426652] env[63175]: DEBUG nova.scheduler.client.report [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 570.462801] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.462801] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.462801] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.463052] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.463052] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.463052] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.463868] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.464438] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.466461] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.466461] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.466461] env[63175]: DEBUG nova.virt.hardware [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.467687] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d18b63c-46a8-486f-8930-ec8829f8db8e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.476741] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aee3dff-2fc7-4e10-a6ca-942a34d0af87 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.938593] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.940567] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 570.943097] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.841s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.046070] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquiring lock "b8b82813-b9c1-4d5c-964f-19016393b985" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.046331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "b8b82813-b9c1-4d5c-964f-19016393b985" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.456423] env[63175]: DEBUG nova.compute.utils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.459459] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 571.459637] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 571.652038] env[63175]: DEBUG nova.policy [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9589dc3047af408d88280afeacf852e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84a2eeb370ba4dc0b5180d2c15f152c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.890968] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d97cd9-3ef5-4314-a37f-dea9bd872f91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.900629] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfeb28e3-a8c9-4aa7-97a3-903621504077 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.935706] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628f0403-6623-478e-b847-83f85f0f673a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.943862] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be75e81b-91b7-4bcc-9b25-d88ecc219d45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.960574] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 571.962519] env[63175]: DEBUG nova.compute.provider_tree [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.466083] env[63175]: DEBUG nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 572.980736] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.981937] env[63175]: ERROR nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Traceback (most recent call last): [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.driver.spawn(context, instance, image_meta, [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] vm_ref = self.build_virtual_machine(instance, [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.981937] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] for vif in network_info: [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self._sync_wrapper(fn, *args, **kwargs) [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.wait() [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self[:] = self._gt.wait() [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self._exit_event.wait() [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] result = hub.switch() [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.982410] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return self.greenlet.switch() [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] result = function(*args, **kwargs) [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] return func(*args, **kwargs) [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise e [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] nwinfo = self.network_api.allocate_for_instance( [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] created_port_ids = self._update_ports_for_instance( [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] with excutils.save_and_reraise_exception(): [ 572.982787] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] self.force_reraise() [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise self.value [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] updated_port = self._update_port( [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] _ensure_no_port_binding_failure(port) [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] raise exception.PortBindingFailed(port_id=port['id']) [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] nova.exception.PortBindingFailed: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. [ 572.983161] env[63175]: ERROR nova.compute.manager [instance: 9595c38c-3699-4b89-bf37-b2e435886878] [ 572.984393] env[63175]: DEBUG nova.compute.utils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.986845] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 572.991535] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Build of instance 9595c38c-3699-4b89-bf37-b2e435886878 was re-scheduled: Binding failed for port c79ee925-f64d-4b19-b0c7-d4b8214c72f9, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 572.992074] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 572.992484] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquiring lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.993873] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Acquired lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.993873] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.995733] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.187s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.998301] env[63175]: DEBUG nova.objects.instance [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 573.030559] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:44:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1991594222',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1500071242',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.030559] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.030559] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.030730] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.030730] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.030730] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.030730] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.032052] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.032052] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.032052] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.032052] env[63175]: DEBUG nova.virt.hardware [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.035305] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10169e16-504a-4d45-859a-fd92efaad3f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.048335] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b94b340-4a22-4cc0-be68-cbfc147d7b7d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.161378] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Successfully created port: 93a34fdb-8933-45b9-a1e2-e4830426e1ef {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.597791] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.012110] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ee4f4e0-ffcb-41d0-970d-b924802547eb tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.013291] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.349s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.289313] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.795676] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Releasing lock "refresh_cache-9595c38c-3699-4b89-bf37-b2e435886878" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.796043] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 574.796138] env[63175]: DEBUG nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.796301] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.804754] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquiring lock "b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.806511] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.858067] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.942106] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec20c96-8216-4144-a513-8e429a58ad5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.952447] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccf6d47-545a-42eb-8cdc-359769070a81 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.995212] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d50ab0e-88b0-44f7-937d-8f42f0e7c0c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.003637] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d2e875-292d-4466-ae71-bb1449cc83c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.020131] env[63175]: DEBUG nova.compute.provider_tree [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.365587] env[63175]: DEBUG nova.network.neutron [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.523514] env[63175]: DEBUG nova.scheduler.client.report [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 575.737410] env[63175]: ERROR nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 575.737410] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.737410] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.737410] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.737410] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.737410] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.737410] env[63175]: ERROR nova.compute.manager raise self.value [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.737410] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.737410] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.737410] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.737983] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.737983] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.737983] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 575.737983] env[63175]: ERROR nova.compute.manager [ 575.737983] env[63175]: Traceback (most recent call last): [ 575.737983] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.737983] env[63175]: listener.cb(fileno) [ 575.737983] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.737983] env[63175]: result = function(*args, **kwargs) [ 575.737983] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.737983] env[63175]: return func(*args, **kwargs) [ 575.737983] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.737983] env[63175]: raise e [ 575.737983] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.737983] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 575.737983] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.737983] env[63175]: created_port_ids = self._update_ports_for_instance( [ 575.737983] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.737983] env[63175]: with excutils.save_and_reraise_exception(): [ 575.737983] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.737983] env[63175]: self.force_reraise() [ 575.737983] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.737983] env[63175]: raise self.value [ 575.737983] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.737983] env[63175]: updated_port = self._update_port( [ 575.737983] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.737983] env[63175]: _ensure_no_port_binding_failure(port) [ 575.737983] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.737983] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.738967] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 575.738967] env[63175]: Removing descriptor: 15 [ 575.738967] env[63175]: ERROR nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Traceback (most recent call last): [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] yield resources [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.driver.spawn(context, instance, image_meta, [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.738967] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] vm_ref = self.build_virtual_machine(instance, [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] for vif in network_info: [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self._sync_wrapper(fn, *args, **kwargs) [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.wait() [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self[:] = self._gt.wait() [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self._exit_event.wait() [ 575.739450] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] result = hub.switch() [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self.greenlet.switch() [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] result = function(*args, **kwargs) [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return func(*args, **kwargs) [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise e [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] nwinfo = self.network_api.allocate_for_instance( [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.739866] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] created_port_ids = self._update_ports_for_instance( [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] with excutils.save_and_reraise_exception(): [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.force_reraise() [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise self.value [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] updated_port = self._update_port( [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] _ensure_no_port_binding_failure(port) [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.740333] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise exception.PortBindingFailed(port_id=port['id']) [ 575.741177] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 575.741177] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] [ 575.741177] env[63175]: INFO nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Terminating instance [ 575.872011] env[63175]: INFO nova.compute.manager [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] [instance: 9595c38c-3699-4b89-bf37-b2e435886878] Took 1.07 seconds to deallocate network for instance. [ 576.031386] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.032640] env[63175]: ERROR nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Traceback (most recent call last): [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.driver.spawn(context, instance, image_meta, [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] vm_ref = self.build_virtual_machine(instance, [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.032640] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] for vif in network_info: [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self._sync_wrapper(fn, *args, **kwargs) [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.wait() [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self[:] = self._gt.wait() [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self._exit_event.wait() [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] result = hub.switch() [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.033105] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return self.greenlet.switch() [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] result = function(*args, **kwargs) [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] return func(*args, **kwargs) [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise e [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] nwinfo = self.network_api.allocate_for_instance( [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] created_port_ids = self._update_ports_for_instance( [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] with excutils.save_and_reraise_exception(): [ 576.033548] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] self.force_reraise() [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise self.value [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] updated_port = self._update_port( [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] _ensure_no_port_binding_failure(port) [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] raise exception.PortBindingFailed(port_id=port['id']) [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] nova.exception.PortBindingFailed: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. [ 576.033941] env[63175]: ERROR nova.compute.manager [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] [ 576.034281] env[63175]: DEBUG nova.compute.utils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.037840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.790s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.042251] env[63175]: INFO nova.compute.claims [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.046876] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Build of instance 596f899b-18f2-42f0-968e-6e9159f1d4d6 was re-scheduled: Binding failed for port 091accda-b313-467a-8f66-b0db11838464, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 576.047454] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 576.048129] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquiring lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.048684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Acquired lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.048684] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.050882] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquiring lock "070514d0-ccc0-4fed-90bd-c7914760db09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.050882] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "070514d0-ccc0-4fed-90bd-c7914760db09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.143354] env[63175]: DEBUG nova.compute.manager [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Received event network-changed-93a34fdb-8933-45b9-a1e2-e4830426e1ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 576.143630] env[63175]: DEBUG nova.compute.manager [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Refreshing instance network info cache due to event network-changed-93a34fdb-8933-45b9-a1e2-e4830426e1ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 576.145700] env[63175]: DEBUG oslo_concurrency.lockutils [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] Acquiring lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.146937] env[63175]: DEBUG oslo_concurrency.lockutils [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] Acquired lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.147209] env[63175]: DEBUG nova.network.neutron [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Refreshing network info cache for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.247602] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquiring lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.550328] env[63175]: ERROR nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 576.550328] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.550328] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.550328] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.550328] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.550328] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.550328] env[63175]: ERROR nova.compute.manager raise self.value [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.550328] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.550328] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.550328] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.551287] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.551287] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.551287] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 576.551287] env[63175]: ERROR nova.compute.manager [ 576.551287] env[63175]: Traceback (most recent call last): [ 576.551287] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.551287] env[63175]: listener.cb(fileno) [ 576.551287] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.551287] env[63175]: result = function(*args, **kwargs) [ 576.551287] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.551287] env[63175]: return func(*args, **kwargs) [ 576.551287] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.551287] env[63175]: raise e [ 576.551287] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.551287] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 576.551287] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.551287] env[63175]: created_port_ids = self._update_ports_for_instance( [ 576.551287] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.551287] env[63175]: with excutils.save_and_reraise_exception(): [ 576.551287] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.551287] env[63175]: self.force_reraise() [ 576.551287] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.551287] env[63175]: raise self.value [ 576.551287] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.551287] env[63175]: updated_port = self._update_port( [ 576.551287] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.551287] env[63175]: _ensure_no_port_binding_failure(port) [ 576.551287] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.551287] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.552918] env[63175]: nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 576.552918] env[63175]: Removing descriptor: 17 [ 576.554995] env[63175]: ERROR nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Traceback (most recent call last): [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] yield resources [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.driver.spawn(context, instance, image_meta, [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] vm_ref = self.build_virtual_machine(instance, [ 576.554995] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] for vif in network_info: [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self._sync_wrapper(fn, *args, **kwargs) [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.wait() [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self[:] = self._gt.wait() [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self._exit_event.wait() [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.555591] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] result = hub.switch() [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self.greenlet.switch() [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] result = function(*args, **kwargs) [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return func(*args, **kwargs) [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise e [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] nwinfo = self.network_api.allocate_for_instance( [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] created_port_ids = self._update_ports_for_instance( [ 576.557642] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] with excutils.save_and_reraise_exception(): [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.force_reraise() [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise self.value [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] updated_port = self._update_port( [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] _ensure_no_port_binding_failure(port) [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise exception.PortBindingFailed(port_id=port['id']) [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 576.558189] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] [ 576.558555] env[63175]: INFO nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Terminating instance [ 576.582917] env[63175]: DEBUG nova.compute.manager [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Received event network-changed-d337dd3a-6f73-421e-b660-f27749e51965 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 576.583134] env[63175]: DEBUG nova.compute.manager [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Refreshing instance network info cache due to event network-changed-d337dd3a-6f73-421e-b660-f27749e51965. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 576.583366] env[63175]: DEBUG oslo_concurrency.lockutils [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] Acquiring lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.583513] env[63175]: DEBUG oslo_concurrency.lockutils [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] Acquired lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.583642] env[63175]: DEBUG nova.network.neutron [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Refreshing network info cache for port d337dd3a-6f73-421e-b660-f27749e51965 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.601702] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.764810] env[63175]: DEBUG nova.network.neutron [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.907131] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.924853] env[63175]: INFO nova.scheduler.client.report [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Deleted allocations for instance 9595c38c-3699-4b89-bf37-b2e435886878 [ 577.067402] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquiring lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.070055] env[63175]: DEBUG nova.network.neutron [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.296792] env[63175]: DEBUG nova.network.neutron [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.412217] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Releasing lock "refresh_cache-596f899b-18f2-42f0-968e-6e9159f1d4d6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.412217] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.412217] env[63175]: DEBUG nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.412217] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.434843] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210df3e3-bd9d-4b18-95d2-bffc9a873e90 tempest-FloatingIPsAssociationTestJSON-860701422 tempest-FloatingIPsAssociationTestJSON-860701422-project-member] Lock "9595c38c-3699-4b89-bf37-b2e435886878" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.728s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.462110] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.485091] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0c1b7c-721f-4642-8850-236c195884c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.496568] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2b4888-e578-42f3-aef8-4a22ed77527f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.533010] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d861eb5-b445-43aa-a698-63a2599f0911 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.546583] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74284032-16cf-4db1-add3-8bba176b10bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.561544] env[63175]: DEBUG nova.compute.provider_tree [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.564943] env[63175]: DEBUG nova.network.neutron [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.574696] env[63175]: DEBUG oslo_concurrency.lockutils [req-0c6c8077-26df-44a2-8cec-dd0775042e06 req-8f2b4635-bf99-4cd0-9464-cfc3e3b7062f service nova] Releasing lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.575212] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquired lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.575296] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.941572] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 577.973623] env[63175]: DEBUG nova.network.neutron [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.067654] env[63175]: DEBUG nova.scheduler.client.report [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 578.073238] env[63175]: DEBUG oslo_concurrency.lockutils [req-d52f0673-4e60-41ea-8dfa-17f486a65c40 req-482e6e94-0fe0-4262-8c03-122f39115c70 service nova] Releasing lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.073238] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquired lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.073238] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.124117] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.396558] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.471019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.480602] env[63175]: INFO nova.compute.manager [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] [instance: 596f899b-18f2-42f0-968e-6e9159f1d4d6] Took 1.07 seconds to deallocate network for instance. [ 578.496886] env[63175]: DEBUG nova.compute.manager [req-1d5a52fb-5866-4e99-9ec1-d89a0f0b0a70 req-f5a5ad06-5be2-44e1-a8a2-3d5bd0c5af40 service nova] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Received event network-vif-deleted-93a34fdb-8933-45b9-a1e2-e4830426e1ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 578.583791] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.584672] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 578.587595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.649s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.590530] env[63175]: INFO nova.compute.claims [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.639100] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.904049] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Releasing lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.904049] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 578.904049] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 578.904419] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a418fcf-fd64-401e-baf7-32a5ee324c64 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.915029] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece74e8a-3f81-48e8-9cd0-4079e29f945a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.937841] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance edd63563-8fe3-4280-9300-8a11ba40ba68 could not be found. [ 578.937841] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 578.937841] env[63175]: INFO nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Took 0.03 seconds to destroy the instance on the hypervisor. [ 578.937841] env[63175]: DEBUG oslo.service.loopingcall [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.938139] env[63175]: DEBUG nova.compute.manager [-] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 578.938139] env[63175]: DEBUG nova.network.neutron [-] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.953225] env[63175]: DEBUG nova.network.neutron [-] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.034931] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.095498] env[63175]: DEBUG nova.compute.utils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.096866] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 579.097721] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.161069] env[63175]: DEBUG nova.compute.manager [req-4d583ddf-2087-49d1-bf9d-11e65f4d5a14 req-2266108d-6982-4ab2-98b3-89963e8e3a83 service nova] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Received event network-vif-deleted-d337dd3a-6f73-421e-b660-f27749e51965 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 579.201679] env[63175]: DEBUG nova.policy [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '663eb8a562ae458da59194b4eb38b8aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c17e0c1c22a94bae96c555868bd5d1f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.460277] env[63175]: DEBUG nova.network.neutron [-] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.531326] env[63175]: INFO nova.scheduler.client.report [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Deleted allocations for instance 596f899b-18f2-42f0-968e-6e9159f1d4d6 [ 579.543266] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Releasing lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.543745] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 579.543935] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.549587] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34410370-1738-4ba5-b6f6-2f708da60930 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.563848] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714e8946-1b7c-4333-9b7e-f3f89f1d96de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.594731] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f33d2617-c4ef-4174-a838-ce0558867b9d could not be found. [ 579.594958] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.595162] env[63175]: INFO nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 579.595418] env[63175]: DEBUG oslo.service.loopingcall [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.595635] env[63175]: DEBUG nova.compute.manager [-] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 579.595723] env[63175]: DEBUG nova.network.neutron [-] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.603148] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 579.665606] env[63175]: DEBUG nova.network.neutron [-] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.889318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquiring lock "a5d426c1-9f30-43a7-942e-06cbda2fce30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.890672] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "a5d426c1-9f30-43a7-942e-06cbda2fce30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.964343] env[63175]: INFO nova.compute.manager [-] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Took 1.02 seconds to deallocate network for instance. [ 579.965576] env[63175]: DEBUG nova.compute.claims [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 579.965758] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.053881] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5da3047e-a27b-48e5-9bc4-18e2b94bb380 tempest-AttachInterfacesUnderV243Test-657592251 tempest-AttachInterfacesUnderV243Test-657592251-project-member] Lock "596f899b-18f2-42f0-968e-6e9159f1d4d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.434s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.117526] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6db8c9-b4d9-4920-a31d-e16fccc7d492 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.129393] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8bc80d-0f49-4c84-8313-ee22e967322f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.173421] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82076f83-25e9-4427-9acd-c5c84b4185be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.176330] env[63175]: DEBUG nova.network.neutron [-] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.182626] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc21a4f9-f2db-415a-8fee-f629cbeaab88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.197288] env[63175]: DEBUG nova.compute.provider_tree [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.356578] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Successfully created port: 7bce0e89-a707-4b90-bccc-54920cb8542a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.556024] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 580.624295] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 580.664268] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.664539] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.664674] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.664863] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.665021] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.666273] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.666273] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.666273] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.666485] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.666599] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.666774] env[63175]: DEBUG nova.virt.hardware [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.667634] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a5cc3a-81d2-47ec-9fd0-9729f4f2b51d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.677106] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88489b3-3458-4ae4-978c-6d1cd15fa305 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.683555] env[63175]: INFO nova.compute.manager [-] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Took 1.09 seconds to deallocate network for instance. [ 580.695571] env[63175]: DEBUG nova.compute.claims [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.695770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.702125] env[63175]: DEBUG nova.scheduler.client.report [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 581.085915] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.207969] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.208404] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 581.214164] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.743s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.214850] env[63175]: INFO nova.compute.claims [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.299441] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "7a25b67c-47b8-420f-9aa2-8c296b8ab221" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.299674] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "7a25b67c-47b8-420f-9aa2-8c296b8ab221" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.719881] env[63175]: DEBUG nova.compute.utils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.727186] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 581.727186] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.813082] env[63175]: DEBUG nova.policy [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ae6287ac33e45adae596ab185c16c16', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '997e36a77dcf4dae890333f724e6945a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.924108] env[63175]: ERROR nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 581.924108] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.924108] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.924108] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.924108] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.924108] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.924108] env[63175]: ERROR nova.compute.manager raise self.value [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.924108] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.924108] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.924108] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.924715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.924715] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.924715] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 581.924715] env[63175]: ERROR nova.compute.manager [ 581.924715] env[63175]: Traceback (most recent call last): [ 581.924715] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.924715] env[63175]: listener.cb(fileno) [ 581.924715] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.924715] env[63175]: result = function(*args, **kwargs) [ 581.924715] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.924715] env[63175]: return func(*args, **kwargs) [ 581.924715] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 581.924715] env[63175]: raise e [ 581.924715] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.924715] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 581.924715] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.924715] env[63175]: created_port_ids = self._update_ports_for_instance( [ 581.924715] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.924715] env[63175]: with excutils.save_and_reraise_exception(): [ 581.924715] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.924715] env[63175]: self.force_reraise() [ 581.924715] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.924715] env[63175]: raise self.value [ 581.924715] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.924715] env[63175]: updated_port = self._update_port( [ 581.924715] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.924715] env[63175]: _ensure_no_port_binding_failure(port) [ 581.924715] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.924715] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.925618] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 581.925618] env[63175]: Removing descriptor: 15 [ 581.925618] env[63175]: ERROR nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Traceback (most recent call last): [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] yield resources [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.driver.spawn(context, instance, image_meta, [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.925618] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] vm_ref = self.build_virtual_machine(instance, [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] for vif in network_info: [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self._sync_wrapper(fn, *args, **kwargs) [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.wait() [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self[:] = self._gt.wait() [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self._exit_event.wait() [ 581.925996] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] result = hub.switch() [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self.greenlet.switch() [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] result = function(*args, **kwargs) [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return func(*args, **kwargs) [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise e [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] nwinfo = self.network_api.allocate_for_instance( [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.926401] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] created_port_ids = self._update_ports_for_instance( [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] with excutils.save_and_reraise_exception(): [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.force_reraise() [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise self.value [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] updated_port = self._update_port( [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] _ensure_no_port_binding_failure(port) [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.926853] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise exception.PortBindingFailed(port_id=port['id']) [ 581.927264] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 581.927264] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] [ 581.927264] env[63175]: INFO nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Terminating instance [ 582.157415] env[63175]: DEBUG nova.compute.manager [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Received event network-changed-7bce0e89-a707-4b90-bccc-54920cb8542a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 582.157695] env[63175]: DEBUG nova.compute.manager [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Refreshing instance network info cache due to event network-changed-7bce0e89-a707-4b90-bccc-54920cb8542a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 582.157976] env[63175]: DEBUG oslo_concurrency.lockutils [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] Acquiring lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.157976] env[63175]: DEBUG oslo_concurrency.lockutils [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] Acquired lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.158298] env[63175]: DEBUG nova.network.neutron [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Refreshing network info cache for port 7bce0e89-a707-4b90-bccc-54920cb8542a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 582.228201] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 582.309170] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Successfully created port: 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.429653] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquiring lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.635340] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb096f33-5ad3-4d5b-ad89-04b4eb06349c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.644758] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685fe775-fee0-4b22-bfda-976f7e272732 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.682518] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1288da72-99f5-49a2-9fce-605c7f386a15 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.692412] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5c1ff4-c3e5-4bac-8795-420180cb471a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.706389] env[63175]: DEBUG nova.compute.provider_tree [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.719542] env[63175]: DEBUG nova.network.neutron [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.790637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquiring lock "d85e98e9-d3d5-4781-8428-d8ab517be146" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.790637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "d85e98e9-d3d5-4781-8428-d8ab517be146" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.930362] env[63175]: DEBUG nova.network.neutron [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.210363] env[63175]: DEBUG nova.scheduler.client.report [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 583.252132] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 583.291460] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.291737] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.291890] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.292084] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.292233] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.292380] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.292594] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.292794] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.292977] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.293154] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.293325] env[63175]: DEBUG nova.virt.hardware [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.294293] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6060205-a105-43c8-8a85-b2d8b23a6016 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.302689] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e49e65-252a-4b7a-9d13-cb349df758b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.435545] env[63175]: DEBUG oslo_concurrency.lockutils [req-368dae38-2402-405d-9f13-1a21d5ff039e req-b274bef2-a7f8-4b95-a723-560c189b68ad service nova] Releasing lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.435545] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquired lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.435545] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.720375] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.721073] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 583.724560] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.859s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.891914] env[63175]: ERROR nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 583.891914] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 583.891914] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.891914] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.891914] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.891914] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.891914] env[63175]: ERROR nova.compute.manager raise self.value [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.891914] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.891914] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.891914] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.892770] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.892770] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.892770] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 583.892770] env[63175]: ERROR nova.compute.manager [ 583.892770] env[63175]: Traceback (most recent call last): [ 583.892770] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.892770] env[63175]: listener.cb(fileno) [ 583.892770] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.892770] env[63175]: result = function(*args, **kwargs) [ 583.892770] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.892770] env[63175]: return func(*args, **kwargs) [ 583.892770] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 583.892770] env[63175]: raise e [ 583.892770] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 583.892770] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 583.892770] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.892770] env[63175]: created_port_ids = self._update_ports_for_instance( [ 583.892770] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.892770] env[63175]: with excutils.save_and_reraise_exception(): [ 583.892770] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.892770] env[63175]: self.force_reraise() [ 583.892770] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.892770] env[63175]: raise self.value [ 583.892770] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.892770] env[63175]: updated_port = self._update_port( [ 583.892770] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.892770] env[63175]: _ensure_no_port_binding_failure(port) [ 583.892770] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.892770] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.895743] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 583.895743] env[63175]: Removing descriptor: 17 [ 583.895743] env[63175]: ERROR nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Traceback (most recent call last): [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] yield resources [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.driver.spawn(context, instance, image_meta, [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.895743] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] vm_ref = self.build_virtual_machine(instance, [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] for vif in network_info: [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self._sync_wrapper(fn, *args, **kwargs) [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.wait() [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self[:] = self._gt.wait() [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self._exit_event.wait() [ 583.896374] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] result = hub.switch() [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self.greenlet.switch() [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] result = function(*args, **kwargs) [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return func(*args, **kwargs) [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise e [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] nwinfo = self.network_api.allocate_for_instance( [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.897189] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] created_port_ids = self._update_ports_for_instance( [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] with excutils.save_and_reraise_exception(): [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.force_reraise() [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise self.value [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] updated_port = self._update_port( [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] _ensure_no_port_binding_failure(port) [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.897956] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise exception.PortBindingFailed(port_id=port['id']) [ 583.898691] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 583.898691] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] [ 583.898691] env[63175]: INFO nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Terminating instance [ 583.930653] env[63175]: DEBUG nova.compute.manager [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Received event network-changed-64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 583.930849] env[63175]: DEBUG nova.compute.manager [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Refreshing instance network info cache due to event network-changed-64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 583.932538] env[63175]: DEBUG oslo_concurrency.lockutils [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] Acquiring lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.932945] env[63175]: DEBUG oslo_concurrency.lockutils [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] Acquired lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.933359] env[63175]: DEBUG nova.network.neutron [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Refreshing network info cache for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.964894] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.082084] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.233707] env[63175]: DEBUG nova.compute.utils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.238054] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 584.242035] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 584.292164] env[63175]: DEBUG nova.compute.manager [req-9ec28dad-23b0-4247-b704-09e39c53fb83 req-f7bc2498-4aa0-45c8-8f91-7977169f6eed service nova] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Received event network-vif-deleted-7bce0e89-a707-4b90-bccc-54920cb8542a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 584.303490] env[63175]: DEBUG nova.policy [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66e12b9981d44ad59a57338b826fdb2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a2474c04b2440fcaf41f74e7e47fb4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.400110] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquiring lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.485631] env[63175]: DEBUG nova.network.neutron [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.585776] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Releasing lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.586190] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 584.588234] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.588234] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ebc7e6b-b407-4e3e-8095-a180101058ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.596508] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42c53b0-b461-45f4-ac81-1579d68ca692 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.608413] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a779f1e-6160-48c6-a387-05b8de4ce10f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.615249] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70896dab-3bda-4dfc-ab3d-9a2cf64575d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.622338] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf could not be found. [ 584.622646] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.622921] env[63175]: INFO nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.623174] env[63175]: DEBUG oslo.service.loopingcall [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.623708] env[63175]: DEBUG nova.compute.manager [-] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 584.623811] env[63175]: DEBUG nova.network.neutron [-] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.655636] env[63175]: DEBUG nova.network.neutron [-] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.657390] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436d11d2-e2b1-4143-b583-63f647a19f2c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.666623] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de16c54-22b4-4e60-a49f-f499d3303173 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.682824] env[63175]: DEBUG nova.compute.provider_tree [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.699039] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Successfully created port: 7d1da76b-3228-4b1b-a259-50d23cc5c021 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.738799] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 584.781787] env[63175]: DEBUG nova.network.neutron [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.161043] env[63175]: DEBUG nova.network.neutron [-] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.186908] env[63175]: DEBUG nova.scheduler.client.report [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 585.283787] env[63175]: DEBUG oslo_concurrency.lockutils [req-30ba51d9-d783-437a-82c5-2806cb2868dc req-a141a89c-46d5-415b-ae51-e1f0469574fa service nova] Releasing lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.284462] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquired lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.284651] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.663561] env[63175]: INFO nova.compute.manager [-] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Took 1.04 seconds to deallocate network for instance. [ 585.665964] env[63175]: DEBUG nova.compute.claims [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.666201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.693263] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.694035] env[63175]: ERROR nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Traceback (most recent call last): [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.driver.spawn(context, instance, image_meta, [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] vm_ref = self.build_virtual_machine(instance, [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.694035] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] for vif in network_info: [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self._sync_wrapper(fn, *args, **kwargs) [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.wait() [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self[:] = self._gt.wait() [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self._exit_event.wait() [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] result = hub.switch() [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.694637] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return self.greenlet.switch() [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] result = function(*args, **kwargs) [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] return func(*args, **kwargs) [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise e [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] nwinfo = self.network_api.allocate_for_instance( [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] created_port_ids = self._update_ports_for_instance( [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] with excutils.save_and_reraise_exception(): [ 585.694994] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] self.force_reraise() [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise self.value [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] updated_port = self._update_port( [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] _ensure_no_port_binding_failure(port) [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] raise exception.PortBindingFailed(port_id=port['id']) [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] nova.exception.PortBindingFailed: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. [ 585.696056] env[63175]: ERROR nova.compute.manager [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] [ 585.696513] env[63175]: DEBUG nova.compute.utils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.696513] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.441s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.696513] env[63175]: DEBUG nova.objects.instance [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 585.698862] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Build of instance e96543c8-dc4e-45b3-8ece-297ee6dbcedb was re-scheduled: Binding failed for port f88926f0-7636-4e1c-b295-ad32eb41e10f, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 585.699389] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 585.699389] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.699645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquired lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.699645] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.748497] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 585.796026] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.796260] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.796626] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.796626] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.796744] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.796858] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.797640] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.797640] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.797769] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.797856] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.798032] env[63175]: DEBUG nova.virt.hardware [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.799258] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de89f2c7-9072-4b31-b47f-a433f98712d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.808912] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7251e6-e9a2-4bee-960b-d7ffd97e3fb4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.828949] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.032950] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.248905] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.292043] env[63175]: DEBUG nova.compute.manager [req-288521c7-f367-4b1e-ac10-e44e7dd9a8f1 req-dacc43d9-b855-41ef-84ed-048d797dd69f service nova] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Received event network-vif-deleted-64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 586.371319] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.536097] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Releasing lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.537726] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 586.538297] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.538297] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be5d06cc-2014-4607-bb05-fc6a115ee407 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.549886] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d04228-941f-4332-ba0b-50947c93d1e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.572866] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d227eb2-b270-4304-91ee-7ae20c33f633 could not be found. [ 586.573118] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.573306] env[63175]: INFO nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Took 0.04 seconds to destroy the instance on the hypervisor. [ 586.573549] env[63175]: DEBUG oslo.service.loopingcall [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.573767] env[63175]: DEBUG nova.compute.manager [-] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.573861] env[63175]: DEBUG nova.network.neutron [-] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.609071] env[63175]: DEBUG nova.network.neutron [-] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.713441] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b527b2b4-df13-4ad8-9515-96914cf66c2d tempest-ServersAdmin275Test-1653102361 tempest-ServersAdmin275Test-1653102361-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.716497] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.275s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.717710] env[63175]: INFO nova.compute.claims [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.868823] env[63175]: ERROR nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 586.868823] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 586.868823] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.868823] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.868823] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.868823] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.868823] env[63175]: ERROR nova.compute.manager raise self.value [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.868823] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.868823] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.868823] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.869413] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.869413] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.869413] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 586.869413] env[63175]: ERROR nova.compute.manager [ 586.869413] env[63175]: Traceback (most recent call last): [ 586.869413] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.869413] env[63175]: listener.cb(fileno) [ 586.869413] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.869413] env[63175]: result = function(*args, **kwargs) [ 586.869413] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.869413] env[63175]: return func(*args, **kwargs) [ 586.869413] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 586.869413] env[63175]: raise e [ 586.869413] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 586.869413] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 586.869413] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.869413] env[63175]: created_port_ids = self._update_ports_for_instance( [ 586.869413] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.869413] env[63175]: with excutils.save_and_reraise_exception(): [ 586.869413] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.869413] env[63175]: self.force_reraise() [ 586.869413] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.869413] env[63175]: raise self.value [ 586.869413] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.869413] env[63175]: updated_port = self._update_port( [ 586.869413] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.869413] env[63175]: _ensure_no_port_binding_failure(port) [ 586.869413] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.869413] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.870228] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 586.870228] env[63175]: Removing descriptor: 17 [ 586.870228] env[63175]: ERROR nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Traceback (most recent call last): [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] yield resources [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.driver.spawn(context, instance, image_meta, [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.870228] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] vm_ref = self.build_virtual_machine(instance, [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] for vif in network_info: [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self._sync_wrapper(fn, *args, **kwargs) [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.wait() [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self[:] = self._gt.wait() [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self._exit_event.wait() [ 586.870592] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] result = hub.switch() [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self.greenlet.switch() [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] result = function(*args, **kwargs) [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return func(*args, **kwargs) [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise e [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] nwinfo = self.network_api.allocate_for_instance( [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.870965] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] created_port_ids = self._update_ports_for_instance( [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] with excutils.save_and_reraise_exception(): [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.force_reraise() [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise self.value [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] updated_port = self._update_port( [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] _ensure_no_port_binding_failure(port) [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.871376] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise exception.PortBindingFailed(port_id=port['id']) [ 586.871924] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 586.871924] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] [ 586.871924] env[63175]: INFO nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Terminating instance [ 586.874007] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Releasing lock "refresh_cache-e96543c8-dc4e-45b3-8ece-297ee6dbcedb" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.874227] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 586.874390] env[63175]: DEBUG nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.874551] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.892257] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.112295] env[63175]: DEBUG nova.network.neutron [-] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.378085] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquiring lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.378354] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquired lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.378447] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.394952] env[63175]: DEBUG nova.network.neutron [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.617902] env[63175]: INFO nova.compute.manager [-] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Took 1.04 seconds to deallocate network for instance. [ 587.620839] env[63175]: DEBUG nova.compute.claims [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.621464] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.897892] env[63175]: INFO nova.compute.manager [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: e96543c8-dc4e-45b3-8ece-297ee6dbcedb] Took 1.02 seconds to deallocate network for instance. [ 587.927843] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.116340] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da817ce5-6b07-4c08-a5f9-6b615dd7f75a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.124428] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b346bdb9-8eff-4a45-a10b-b862e57ef4a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.157442] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.162018] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76689ca-9643-454a-9622-c1cf5cb04038 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.167375] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d893abf4-a32a-44be-be48-0888431487b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.182022] env[63175]: DEBUG nova.compute.provider_tree [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.367277] env[63175]: DEBUG nova.compute.manager [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Received event network-changed-7d1da76b-3228-4b1b-a259-50d23cc5c021 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 588.367490] env[63175]: DEBUG nova.compute.manager [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Refreshing instance network info cache due to event network-changed-7d1da76b-3228-4b1b-a259-50d23cc5c021. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 588.367674] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] Acquiring lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.663961] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Releasing lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.664294] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 588.664294] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.664558] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] Acquired lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.664765] env[63175]: DEBUG nova.network.neutron [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Refreshing network info cache for port 7d1da76b-3228-4b1b-a259-50d23cc5c021 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.666089] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1a631fc-851e-4749-ba46-eba95b7fdbae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.685343] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b617025-897b-44c8-90cb-4f5fa4619de2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.701355] env[63175]: DEBUG nova.scheduler.client.report [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 588.720042] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c50a11c-cab7-461f-9550-575c4f51a4e5 could not be found. [ 588.720042] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.720211] env[63175]: INFO nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 588.720492] env[63175]: DEBUG oslo.service.loopingcall [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.720801] env[63175]: DEBUG nova.compute.manager [-] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 588.720944] env[63175]: DEBUG nova.network.neutron [-] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.918062] env[63175]: DEBUG nova.network.neutron [-] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.946348] env[63175]: INFO nova.scheduler.client.report [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Deleted allocations for instance e96543c8-dc4e-45b3-8ece-297ee6dbcedb [ 589.202516] env[63175]: DEBUG nova.network.neutron [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.207872] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.208388] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 589.211093] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.770s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.213077] env[63175]: INFO nova.compute.claims [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.377389] env[63175]: DEBUG nova.network.neutron [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.420729] env[63175]: DEBUG nova.network.neutron [-] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.456933] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bc2a39bb-c844-4426-867c-b4c4ae1c9bf4 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "e96543c8-dc4e-45b3-8ece-297ee6dbcedb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.979s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.723543] env[63175]: DEBUG nova.compute.utils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.725798] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 589.726301] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 589.831200] env[63175]: DEBUG nova.policy [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f41d93d6fff498f86c024150d9cc8ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cddeb1059c942f2b603ffbb7a91ecff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.883655] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] Releasing lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.883923] env[63175]: DEBUG nova.compute.manager [req-8f3b728d-6c33-4a74-84d7-e88515d4dde5 req-8b24fe16-36c6-442b-8830-d2267a23498e service nova] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Received event network-vif-deleted-7d1da76b-3228-4b1b-a259-50d23cc5c021 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 589.923687] env[63175]: INFO nova.compute.manager [-] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Took 1.20 seconds to deallocate network for instance. [ 589.928322] env[63175]: DEBUG nova.compute.claims [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.928322] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.960526] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 590.160078] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "11747695-c79c-477d-b5ae-44c49dfb4bba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.160078] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "11747695-c79c-477d-b5ae-44c49dfb4bba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.204104] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "8b305bcb-87e0-4fc9-9579-56a5af9ace8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.204104] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "8b305bcb-87e0-4fc9-9579-56a5af9ace8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.236036] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 590.327253] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Successfully created port: b8256eb0-ef2d-49ca-9ed4-c54312c9505a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.509041] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.792997] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ebfdea-4f68-4a17-a723-c04fa67f8da5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.800398] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "182b1cc6-9626-4146-be9a-e15c5690aa1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.800963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "182b1cc6-9626-4146-be9a-e15c5690aa1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.805970] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecde11f-4bfd-4a3e-8926-65f8f59a779f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.854461] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c48dd1-8205-4194-bc26-ae96b6e7b512 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.863233] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e654d8c2-147e-4df5-88f3-dfa2051f1586 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.882325] env[63175]: DEBUG nova.compute.provider_tree [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.262283] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 591.286458] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.286458] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.286458] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.286667] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.286667] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.286667] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.286667] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.286667] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.287175] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.287563] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.288470] env[63175]: DEBUG nova.virt.hardware [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.289151] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb12497-7763-4472-9795-d6d2cbaa496b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.301486] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3ffb19-f06d-4e85-ae9e-94e7d477c2c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.387019] env[63175]: DEBUG nova.scheduler.client.report [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 591.425230] env[63175]: DEBUG nova.compute.manager [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Received event network-changed-b8256eb0-ef2d-49ca-9ed4-c54312c9505a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 591.425430] env[63175]: DEBUG nova.compute.manager [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Refreshing instance network info cache due to event network-changed-b8256eb0-ef2d-49ca-9ed4-c54312c9505a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 591.425642] env[63175]: DEBUG oslo_concurrency.lockutils [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] Acquiring lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.425776] env[63175]: DEBUG oslo_concurrency.lockutils [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] Acquired lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.425935] env[63175]: DEBUG nova.network.neutron [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Refreshing network info cache for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.590360] env[63175]: ERROR nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 591.590360] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 591.590360] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.590360] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.590360] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.590360] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.590360] env[63175]: ERROR nova.compute.manager raise self.value [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.590360] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.590360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.590360] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.590940] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.590940] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.590940] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 591.590940] env[63175]: ERROR nova.compute.manager [ 591.590940] env[63175]: Traceback (most recent call last): [ 591.590940] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.590940] env[63175]: listener.cb(fileno) [ 591.590940] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.590940] env[63175]: result = function(*args, **kwargs) [ 591.590940] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.590940] env[63175]: return func(*args, **kwargs) [ 591.590940] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 591.590940] env[63175]: raise e [ 591.590940] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 591.590940] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 591.590940] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.590940] env[63175]: created_port_ids = self._update_ports_for_instance( [ 591.590940] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.590940] env[63175]: with excutils.save_and_reraise_exception(): [ 591.590940] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.590940] env[63175]: self.force_reraise() [ 591.590940] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.590940] env[63175]: raise self.value [ 591.590940] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.590940] env[63175]: updated_port = self._update_port( [ 591.590940] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.590940] env[63175]: _ensure_no_port_binding_failure(port) [ 591.590940] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.590940] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.591789] env[63175]: nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 591.591789] env[63175]: Removing descriptor: 17 [ 591.591789] env[63175]: ERROR nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Traceback (most recent call last): [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] yield resources [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.driver.spawn(context, instance, image_meta, [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.591789] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] vm_ref = self.build_virtual_machine(instance, [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] for vif in network_info: [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self._sync_wrapper(fn, *args, **kwargs) [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.wait() [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self[:] = self._gt.wait() [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self._exit_event.wait() [ 591.592358] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] result = hub.switch() [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self.greenlet.switch() [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] result = function(*args, **kwargs) [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return func(*args, **kwargs) [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise e [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] nwinfo = self.network_api.allocate_for_instance( [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.592731] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] created_port_ids = self._update_ports_for_instance( [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] with excutils.save_and_reraise_exception(): [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.force_reraise() [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise self.value [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] updated_port = self._update_port( [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] _ensure_no_port_binding_failure(port) [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.593431] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise exception.PortBindingFailed(port_id=port['id']) [ 591.593940] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 591.593940] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] [ 591.593940] env[63175]: INFO nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Terminating instance [ 591.894170] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.894170] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 591.896191] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.370s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.896974] env[63175]: DEBUG nova.objects.instance [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lazy-loading 'resources' on Instance uuid 56b3ba41-0406-4b69-89af-b00e87feaa89 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 591.951675] env[63175]: DEBUG nova.network.neutron [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.023412] env[63175]: DEBUG nova.network.neutron [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.095920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquiring lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.399951] env[63175]: DEBUG nova.compute.utils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.401407] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 592.401489] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 592.417930] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "8aaf9c9f-e992-49ca-acad-d518503544e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.418338] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "8aaf9c9f-e992-49ca-acad-d518503544e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.459242] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "223ccdbc-cee3-4040-a1ee-a53929025926" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.459813] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "223ccdbc-cee3-4040-a1ee-a53929025926" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.483714] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "d5bc39a7-a493-41a3-a2f4-8958883adca8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.484015] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "d5bc39a7-a493-41a3-a2f4-8958883adca8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.491368] env[63175]: DEBUG nova.policy [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.525218] env[63175]: DEBUG oslo_concurrency.lockutils [req-0da9cafb-f99c-45b4-9812-6548ccb55b39 req-1388f72d-2e1a-47a8-8bdb-d89450de6edc service nova] Releasing lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.526894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquired lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.526894] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.836339] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Successfully created port: 051a669a-01ac-4efc-8711-1e76c6cac8a8 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.896036] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee901eb-6d90-4166-b413-5559bc89bee2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.904139] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcee734-c299-4c3d-9f96-3a7367207b0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.944507] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 592.948055] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1f7c3e-0090-41c9-b250-beb9e40c2a17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.958279] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6091de-7737-4413-84de-0036ecedc58f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.969746] env[63175]: DEBUG nova.compute.provider_tree [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.058289] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.140875] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.321098] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "9af5d212-c032-4446-b3cd-87fe4e66ad7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.321098] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "9af5d212-c032-4446-b3cd-87fe4e66ad7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.473033] env[63175]: DEBUG nova.scheduler.client.report [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 593.569293] env[63175]: DEBUG nova.compute.manager [req-755416ad-377f-4317-a89a-ef02026e67e9 req-7eff578d-8b12-4f1e-9f39-d8f734ee8517 service nova] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Received event network-vif-deleted-b8256eb0-ef2d-49ca-9ed4-c54312c9505a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 593.645595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Releasing lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.646402] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 593.646402] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.646402] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5415af63-8ece-4969-baeb-7bfad7e45021 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.656377] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835777b0-0fa2-495e-8432-eae98660b853 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.681909] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 652f4c1b-69b4-415d-ab14-bbf10043471d could not be found. [ 593.684019] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.684019] env[63175]: INFO nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.684019] env[63175]: DEBUG oslo.service.loopingcall [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.684019] env[63175]: DEBUG nova.compute.manager [-] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 593.684019] env[63175]: DEBUG nova.network.neutron [-] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.698628] env[63175]: DEBUG nova.network.neutron [-] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.966522] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 593.981094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.082s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.981094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.657s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.981960] env[63175]: DEBUG nova.objects.instance [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lazy-loading 'resources' on Instance uuid 24ee74d8-33d2-4efa-97de-a642c89a3461 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 594.005860] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.005860] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.006078] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.006398] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.006500] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.006647] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.006857] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.007036] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.007691] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.008033] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.008413] env[63175]: DEBUG nova.virt.hardware [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.010330] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22945446-332c-4a26-a1fb-b3dcdc9dd7fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.014551] env[63175]: INFO nova.scheduler.client.report [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Deleted allocations for instance 56b3ba41-0406-4b69-89af-b00e87feaa89 [ 594.022505] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d220a86-42e1-40ad-a200-d5b625bc32f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.044844] env[63175]: ERROR nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 594.044844] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.044844] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.044844] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.044844] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.044844] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.044844] env[63175]: ERROR nova.compute.manager raise self.value [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.044844] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.044844] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.044844] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.045362] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.045362] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.045362] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 594.045362] env[63175]: ERROR nova.compute.manager [ 594.045362] env[63175]: Traceback (most recent call last): [ 594.045362] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.045362] env[63175]: listener.cb(fileno) [ 594.045362] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.045362] env[63175]: result = function(*args, **kwargs) [ 594.045362] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.045362] env[63175]: return func(*args, **kwargs) [ 594.045362] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.045362] env[63175]: raise e [ 594.045362] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.045362] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 594.045362] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.045362] env[63175]: created_port_ids = self._update_ports_for_instance( [ 594.045362] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.045362] env[63175]: with excutils.save_and_reraise_exception(): [ 594.045362] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.045362] env[63175]: self.force_reraise() [ 594.045362] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.045362] env[63175]: raise self.value [ 594.045362] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.045362] env[63175]: updated_port = self._update_port( [ 594.045362] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.045362] env[63175]: _ensure_no_port_binding_failure(port) [ 594.045362] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.045362] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.046171] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 594.046171] env[63175]: Removing descriptor: 17 [ 594.046171] env[63175]: ERROR nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Traceback (most recent call last): [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] yield resources [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.driver.spawn(context, instance, image_meta, [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.046171] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] vm_ref = self.build_virtual_machine(instance, [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] for vif in network_info: [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self._sync_wrapper(fn, *args, **kwargs) [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.wait() [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self[:] = self._gt.wait() [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self._exit_event.wait() [ 594.046657] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] result = hub.switch() [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self.greenlet.switch() [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] result = function(*args, **kwargs) [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return func(*args, **kwargs) [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise e [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] nwinfo = self.network_api.allocate_for_instance( [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.047040] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] created_port_ids = self._update_ports_for_instance( [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] with excutils.save_and_reraise_exception(): [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.force_reraise() [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise self.value [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] updated_port = self._update_port( [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] _ensure_no_port_binding_failure(port) [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.047418] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise exception.PortBindingFailed(port_id=port['id']) [ 594.047754] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 594.047754] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] [ 594.047754] env[63175]: INFO nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Terminating instance [ 594.201362] env[63175]: DEBUG nova.network.neutron [-] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.533244] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7af07cd9-1c5d-4a91-a96b-4acf53f6fe3d tempest-ServerDiagnosticsV248Test-1995185387 tempest-ServerDiagnosticsV248Test-1995185387-project-member] Lock "56b3ba41-0406-4b69-89af-b00e87feaa89" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.490s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.554233] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.556534] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.556534] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.705782] env[63175]: INFO nova.compute.manager [-] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Took 1.02 seconds to deallocate network for instance. [ 594.709228] env[63175]: DEBUG nova.compute.claims [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.709228] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.963058] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178f078b-1389-455a-a3f8-8d35323ca790 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.971333] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d01716-955b-425b-824c-a7592694c842 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.004607] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c49fff1-deea-4000-bc6c-1572f8172dec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.012717] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b466c0c5-b19f-4131-877f-6b783551abd6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.026710] env[63175]: DEBUG nova.compute.provider_tree [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.080707] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.208125] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquiring lock "295e958d-20d4-493c-856b-71880cce8b98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.208426] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "295e958d-20d4-493c-856b-71880cce8b98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.222886] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.529860] env[63175]: DEBUG nova.scheduler.client.report [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 595.592057] env[63175]: DEBUG nova.compute.manager [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Received event network-changed-051a669a-01ac-4efc-8711-1e76c6cac8a8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 595.592265] env[63175]: DEBUG nova.compute.manager [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Refreshing instance network info cache due to event network-changed-051a669a-01ac-4efc-8711-1e76c6cac8a8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 595.592474] env[63175]: DEBUG oslo_concurrency.lockutils [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] Acquiring lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.727059] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.727604] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 595.727882] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 595.728268] env[63175]: DEBUG oslo_concurrency.lockutils [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] Acquired lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.728449] env[63175]: DEBUG nova.network.neutron [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Refreshing network info cache for port 051a669a-01ac-4efc-8711-1e76c6cac8a8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.729525] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-190b8ee8-1163-4a79-91b8-545a9ff9045e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.740185] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce86faf-917a-4d21-9e49-3bb3f50fd438 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.764647] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5 could not be found. [ 595.764899] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.765125] env[63175]: INFO nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.765325] env[63175]: DEBUG oslo.service.loopingcall [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.765546] env[63175]: DEBUG nova.compute.manager [-] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 595.765642] env[63175]: DEBUG nova.network.neutron [-] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.780418] env[63175]: DEBUG nova.network.neutron [-] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.036364] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.038645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.568s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.040287] env[63175]: INFO nova.compute.claims [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.061876] env[63175]: INFO nova.scheduler.client.report [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Deleted allocations for instance 24ee74d8-33d2-4efa-97de-a642c89a3461 [ 596.249627] env[63175]: DEBUG nova.network.neutron [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.283917] env[63175]: DEBUG nova.network.neutron [-] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.333121] env[63175]: DEBUG nova.network.neutron [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.569726] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4370330-f852-442c-89a2-dd52a45b2517 tempest-ServersAdmin275Test-1882975909 tempest-ServersAdmin275Test-1882975909-project-member] Lock "24ee74d8-33d2-4efa-97de-a642c89a3461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.834s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.715789] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquiring lock "4dcb8813-c428-4e0f-a761-3b06a8cba173" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.716040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "4dcb8813-c428-4e0f-a761-3b06a8cba173" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.786023] env[63175]: INFO nova.compute.manager [-] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Took 1.02 seconds to deallocate network for instance. [ 596.788476] env[63175]: DEBUG nova.compute.claims [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.788656] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.835588] env[63175]: DEBUG oslo_concurrency.lockutils [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] Releasing lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.835858] env[63175]: DEBUG nova.compute.manager [req-ac612528-367c-4f5c-8df7-1a47f617d582 req-359a09d9-b683-4143-b087-08cad1475738 service nova] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Received event network-vif-deleted-051a669a-01ac-4efc-8711-1e76c6cac8a8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 597.447145] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141bccfc-b7a3-4ed5-ac5b-64afd80f0818 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.453832] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2d342f-31b8-4840-8900-844cc574903a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.488436] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8545467-3c19-4e3a-bc1a-991582356b75 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.496846] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25999ba7-7de0-4e71-b06e-81dd8399fee8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.510620] env[63175]: DEBUG nova.compute.provider_tree [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.015408] env[63175]: DEBUG nova.scheduler.client.report [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 598.521471] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.522061] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 598.525086] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.559s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.032017] env[63175]: DEBUG nova.compute.utils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.035845] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 599.036000] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.119610] env[63175]: DEBUG nova.policy [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5ae88db36b14c63a4256b99d607ee8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b2bd3d977bf4865baa2c8bb4522349e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.443158] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c54d912-46e2-4c12-889b-02cdebc89d35 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.452302] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db6ef0b-965e-4363-b9eb-4e6bc6059dd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.491350] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96266e18-ee5d-47a9-a726-51d39aa9a500 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.499785] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e57d6e-e37a-41c9-b750-e507ab9709b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.514631] env[63175]: DEBUG nova.compute.provider_tree [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.536073] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 599.544109] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Successfully created port: 71cf3fb8-b22b-4285-a520-b7a96bdd17bc {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.669152] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquiring lock "0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.669698] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.018736] env[63175]: DEBUG nova.scheduler.client.report [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 600.513464] env[63175]: DEBUG nova.compute.manager [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Received event network-changed-71cf3fb8-b22b-4285-a520-b7a96bdd17bc {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 600.513464] env[63175]: DEBUG nova.compute.manager [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Refreshing instance network info cache due to event network-changed-71cf3fb8-b22b-4285-a520-b7a96bdd17bc. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 600.514964] env[63175]: DEBUG oslo_concurrency.lockutils [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] Acquiring lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.515245] env[63175]: DEBUG oslo_concurrency.lockutils [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] Acquired lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.519336] env[63175]: DEBUG nova.network.neutron [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Refreshing network info cache for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.523805] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.525686] env[63175]: ERROR nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Traceback (most recent call last): [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.driver.spawn(context, instance, image_meta, [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] vm_ref = self.build_virtual_machine(instance, [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.525686] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] for vif in network_info: [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self._sync_wrapper(fn, *args, **kwargs) [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.wait() [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self[:] = self._gt.wait() [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self._exit_event.wait() [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] result = hub.switch() [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.526750] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return self.greenlet.switch() [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] result = function(*args, **kwargs) [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] return func(*args, **kwargs) [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise e [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] nwinfo = self.network_api.allocate_for_instance( [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] created_port_ids = self._update_ports_for_instance( [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] with excutils.save_and_reraise_exception(): [ 600.527583] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] self.force_reraise() [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise self.value [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] updated_port = self._update_port( [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] _ensure_no_port_binding_failure(port) [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] raise exception.PortBindingFailed(port_id=port['id']) [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] nova.exception.PortBindingFailed: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. [ 600.528234] env[63175]: ERROR nova.compute.manager [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] [ 600.528630] env[63175]: DEBUG nova.compute.utils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.528630] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.832s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.531867] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Build of instance edd63563-8fe3-4280-9300-8a11ba40ba68 was re-scheduled: Binding failed for port 93a34fdb-8933-45b9-a1e2-e4830426e1ef, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 600.532358] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 600.532578] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquiring lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.532725] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Acquired lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.532915] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.545136] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 600.570379] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.570634] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.570789] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.570971] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.571133] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.571278] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.571891] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.571891] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.571891] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.572118] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.572118] env[63175]: DEBUG nova.virt.hardware [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.572949] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db563fd-c7e9-4ded-8e9b-47e3492cc305 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.582779] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbb849b-408b-4ba7-8e05-e23ad918f79c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.739715] env[63175]: ERROR nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 600.739715] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 600.739715] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.739715] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.739715] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.739715] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.739715] env[63175]: ERROR nova.compute.manager raise self.value [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.739715] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.739715] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.739715] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.740187] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.740187] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.740187] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 600.740187] env[63175]: ERROR nova.compute.manager [ 600.740187] env[63175]: Traceback (most recent call last): [ 600.740187] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.740187] env[63175]: listener.cb(fileno) [ 600.740187] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.740187] env[63175]: result = function(*args, **kwargs) [ 600.740187] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.740187] env[63175]: return func(*args, **kwargs) [ 600.740187] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 600.740187] env[63175]: raise e [ 600.740187] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 600.740187] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 600.740187] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.740187] env[63175]: created_port_ids = self._update_ports_for_instance( [ 600.740187] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.740187] env[63175]: with excutils.save_and_reraise_exception(): [ 600.740187] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.740187] env[63175]: self.force_reraise() [ 600.740187] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.740187] env[63175]: raise self.value [ 600.740187] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.740187] env[63175]: updated_port = self._update_port( [ 600.740187] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.740187] env[63175]: _ensure_no_port_binding_failure(port) [ 600.740187] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.740187] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.740959] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 600.740959] env[63175]: Removing descriptor: 17 [ 600.740959] env[63175]: ERROR nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Traceback (most recent call last): [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] yield resources [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.driver.spawn(context, instance, image_meta, [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.740959] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] vm_ref = self.build_virtual_machine(instance, [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] for vif in network_info: [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self._sync_wrapper(fn, *args, **kwargs) [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.wait() [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self[:] = self._gt.wait() [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self._exit_event.wait() [ 600.741212] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] result = hub.switch() [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self.greenlet.switch() [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] result = function(*args, **kwargs) [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return func(*args, **kwargs) [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise e [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] nwinfo = self.network_api.allocate_for_instance( [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.741527] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] created_port_ids = self._update_ports_for_instance( [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] with excutils.save_and_reraise_exception(): [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.force_reraise() [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise self.value [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] updated_port = self._update_port( [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] _ensure_no_port_binding_failure(port) [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.741784] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise exception.PortBindingFailed(port_id=port['id']) [ 600.742072] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 600.742072] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] [ 600.742072] env[63175]: INFO nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Terminating instance [ 601.041010] env[63175]: DEBUG nova.network.neutron [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.058431] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.107029] env[63175]: DEBUG nova.network.neutron [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.183411] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.250025] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.554507] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4eb3f04-e462-4d68-b907-402c38eb9f17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.562732] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2d479c-236b-4b37-8bbd-bb117e8e0c3a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.602213] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6838cc0-f1be-4f54-a9db-7c1a197d468d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.609014] env[63175]: DEBUG oslo_concurrency.lockutils [req-dc36d98c-d4ee-4e50-b917-edaebd2bed7b req-560f466f-f1b0-4851-b1cc-461daad2986b service nova] Releasing lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.609490] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquired lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.609688] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.611922] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898b9bda-8709-45ec-b2ee-cf32a75695af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.625555] env[63175]: DEBUG nova.compute.provider_tree [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.691940] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Releasing lock "refresh_cache-edd63563-8fe3-4280-9300-8a11ba40ba68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.692584] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 601.692584] env[63175]: DEBUG nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 601.692867] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.718471] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.128269] env[63175]: DEBUG nova.scheduler.client.report [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 602.135570] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.188577] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.221230] env[63175]: DEBUG nova.network.neutron [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.563663] env[63175]: DEBUG nova.compute.manager [req-7056bca9-f507-434e-a8a5-93bfd661b467 req-dfcea9ba-4afd-484b-bca5-1722aa8e42af service nova] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Received event network-vif-deleted-71cf3fb8-b22b-4285-a520-b7a96bdd17bc {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 602.639383] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.111s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.640061] env[63175]: ERROR nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Traceback (most recent call last): [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.driver.spawn(context, instance, image_meta, [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] vm_ref = self.build_virtual_machine(instance, [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.640061] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] for vif in network_info: [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self._sync_wrapper(fn, *args, **kwargs) [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.wait() [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self[:] = self._gt.wait() [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self._exit_event.wait() [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] result = hub.switch() [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.640399] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return self.greenlet.switch() [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] result = function(*args, **kwargs) [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] return func(*args, **kwargs) [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise e [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] nwinfo = self.network_api.allocate_for_instance( [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] created_port_ids = self._update_ports_for_instance( [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] with excutils.save_and_reraise_exception(): [ 602.640746] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] self.force_reraise() [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise self.value [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] updated_port = self._update_port( [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] _ensure_no_port_binding_failure(port) [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] raise exception.PortBindingFailed(port_id=port['id']) [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] nova.exception.PortBindingFailed: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. [ 602.641040] env[63175]: ERROR nova.compute.manager [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] [ 602.641300] env[63175]: DEBUG nova.compute.utils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.642021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.556s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.643539] env[63175]: INFO nova.compute.claims [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.651082] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Build of instance f33d2617-c4ef-4174-a838-ce0558867b9d was re-scheduled: Binding failed for port d337dd3a-6f73-421e-b660-f27749e51965, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 602.651082] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 602.651082] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquiring lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.651316] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Acquired lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.651316] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.692784] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Releasing lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.694227] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 602.694227] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.694883] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ae7f9a4-be30-4cbf-9fcc-166e9cb110c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.709914] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e144ea7-c224-4443-acfc-714deeb94cb5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.728988] env[63175]: INFO nova.compute.manager [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] [instance: edd63563-8fe3-4280-9300-8a11ba40ba68] Took 1.04 seconds to deallocate network for instance. [ 602.744349] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4883dd7f-a13e-4efc-90ff-ca42909a7f0c could not be found. [ 602.744581] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.744760] env[63175]: INFO nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 602.744999] env[63175]: DEBUG oslo.service.loopingcall [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.745497] env[63175]: DEBUG nova.compute.manager [-] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 602.745597] env[63175]: DEBUG nova.network.neutron [-] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.772238] env[63175]: DEBUG nova.network.neutron [-] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.176140] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.258461] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.278055] env[63175]: DEBUG nova.network.neutron [-] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.760944] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Releasing lock "refresh_cache-f33d2617-c4ef-4174-a838-ce0558867b9d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.761204] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 603.761379] env[63175]: DEBUG nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.761539] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.775584] env[63175]: INFO nova.scheduler.client.report [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Deleted allocations for instance edd63563-8fe3-4280-9300-8a11ba40ba68 [ 603.782258] env[63175]: INFO nova.compute.manager [-] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Took 1.04 seconds to deallocate network for instance. [ 603.793242] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.794259] env[63175]: DEBUG nova.compute.claims [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.794259] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.153744] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037c2bd9-0c11-4e41-9ec4-640c32583261 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.161540] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6e1308-fbe8-44bf-937f-4b96b4dc90ed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.192166] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e733cc80-4ba9-4ffd-9014-e2cbe82bdd82 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.199147] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b07d53-b01c-4836-b568-5f66e3afb2cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.829601] env[63175]: DEBUG nova.network.neutron [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.830934] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b3a911d-de87-455d-97fa-730e81195d5c tempest-ServersWithSpecificFlavorTestJSON-903303062 tempest-ServersWithSpecificFlavorTestJSON-903303062-project-member] Lock "edd63563-8fe3-4280-9300-8a11ba40ba68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.009s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.846334] env[63175]: DEBUG nova.compute.provider_tree [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.847966] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 605.333239] env[63175]: INFO nova.compute.manager [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] [instance: f33d2617-c4ef-4174-a838-ce0558867b9d] Took 1.57 seconds to deallocate network for instance. [ 605.352212] env[63175]: DEBUG nova.scheduler.client.report [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 605.376860] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.859302] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.217s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.859972] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 605.863618] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.197s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.371742] env[63175]: DEBUG nova.compute.utils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.381268] env[63175]: INFO nova.scheduler.client.report [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Deleted allocations for instance f33d2617-c4ef-4174-a838-ce0558867b9d [ 606.387133] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 606.387520] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 606.463787] env[63175]: DEBUG nova.policy [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5ae88db36b14c63a4256b99d607ee8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b2bd3d977bf4865baa2c8bb4522349e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.827391] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Successfully created port: f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.841739] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dc5623-22fd-4240-b6fe-8979e1377831 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.853533] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7309f89a-1875-4cc1-8cac-c85fbe9e3c71 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.888757] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 606.892190] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f8738ac7-e6ae-4ee5-9e50-1668e1dac1c4 tempest-InstanceActionsTestJSON-669739213 tempest-InstanceActionsTestJSON-669739213-project-member] Lock "f33d2617-c4ef-4174-a838-ce0558867b9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.062s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.892560] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a235179-8c59-4e8c-bdc2-fbe23ecb1fc2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.908791] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a703b9e3-ec2e-4bc7-b556-6c65de055936 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.934086] env[63175]: DEBUG nova.compute.provider_tree [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.401506] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 607.438267] env[63175]: DEBUG nova.scheduler.client.report [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 607.909164] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 607.939205] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.941825] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.941981] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.942154] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.942348] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.942491] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.942638] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.942863] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.943149] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.943365] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.943556] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.943702] env[63175]: DEBUG nova.virt.hardware [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.944575] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e44050d-8010-4a82-a611-0701bb7c6461 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.947896] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.084s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.948727] env[63175]: ERROR nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Traceback (most recent call last): [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.driver.spawn(context, instance, image_meta, [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] vm_ref = self.build_virtual_machine(instance, [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.948727] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] for vif in network_info: [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self._sync_wrapper(fn, *args, **kwargs) [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.wait() [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self[:] = self._gt.wait() [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self._exit_event.wait() [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] result = hub.switch() [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.949079] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return self.greenlet.switch() [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] result = function(*args, **kwargs) [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] return func(*args, **kwargs) [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise e [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] nwinfo = self.network_api.allocate_for_instance( [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] created_port_ids = self._update_ports_for_instance( [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] with excutils.save_and_reraise_exception(): [ 607.949346] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] self.force_reraise() [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise self.value [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] updated_port = self._update_port( [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] _ensure_no_port_binding_failure(port) [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] raise exception.PortBindingFailed(port_id=port['id']) [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] nova.exception.PortBindingFailed: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. [ 607.949676] env[63175]: ERROR nova.compute.manager [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] [ 607.949912] env[63175]: DEBUG nova.compute.utils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.951542] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.331s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.955656] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Build of instance 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf was re-scheduled: Binding failed for port 7bce0e89-a707-4b90-bccc-54920cb8542a, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 607.956123] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 607.956353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquiring lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.956498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Acquired lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.956654] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.961871] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee6d634-e2c5-4250-ae37-f5571d7377e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.041809] env[63175]: DEBUG nova.compute.manager [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Received event network-changed-f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 608.042016] env[63175]: DEBUG nova.compute.manager [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Refreshing instance network info cache due to event network-changed-f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 608.042245] env[63175]: DEBUG oslo_concurrency.lockutils [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] Acquiring lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.042383] env[63175]: DEBUG oslo_concurrency.lockutils [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] Acquired lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.042537] env[63175]: DEBUG nova.network.neutron [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Refreshing network info cache for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.351707] env[63175]: ERROR nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 608.351707] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 608.351707] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.351707] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.351707] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.351707] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.351707] env[63175]: ERROR nova.compute.manager raise self.value [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.351707] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.351707] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.351707] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.352123] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.352123] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.352123] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 608.352123] env[63175]: ERROR nova.compute.manager [ 608.352123] env[63175]: Traceback (most recent call last): [ 608.352123] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.352123] env[63175]: listener.cb(fileno) [ 608.352123] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.352123] env[63175]: result = function(*args, **kwargs) [ 608.352123] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.352123] env[63175]: return func(*args, **kwargs) [ 608.352123] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 608.352123] env[63175]: raise e [ 608.352123] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 608.352123] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 608.352123] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.352123] env[63175]: created_port_ids = self._update_ports_for_instance( [ 608.352123] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.352123] env[63175]: with excutils.save_and_reraise_exception(): [ 608.352123] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.352123] env[63175]: self.force_reraise() [ 608.352123] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.352123] env[63175]: raise self.value [ 608.352123] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.352123] env[63175]: updated_port = self._update_port( [ 608.352123] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.352123] env[63175]: _ensure_no_port_binding_failure(port) [ 608.352123] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.352123] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.352706] env[63175]: nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 608.352706] env[63175]: Removing descriptor: 15 [ 608.352706] env[63175]: ERROR nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Traceback (most recent call last): [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] yield resources [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.driver.spawn(context, instance, image_meta, [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.352706] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] vm_ref = self.build_virtual_machine(instance, [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] for vif in network_info: [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self._sync_wrapper(fn, *args, **kwargs) [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.wait() [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self[:] = self._gt.wait() [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self._exit_event.wait() [ 608.353017] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] result = hub.switch() [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self.greenlet.switch() [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] result = function(*args, **kwargs) [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return func(*args, **kwargs) [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise e [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] nwinfo = self.network_api.allocate_for_instance( [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.353290] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] created_port_ids = self._update_ports_for_instance( [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] with excutils.save_and_reraise_exception(): [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.force_reraise() [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise self.value [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] updated_port = self._update_port( [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] _ensure_no_port_binding_failure(port) [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.353607] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise exception.PortBindingFailed(port_id=port['id']) [ 608.353858] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 608.353858] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] [ 608.353858] env[63175]: INFO nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Terminating instance [ 608.491508] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.583488] env[63175]: DEBUG nova.network.neutron [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.613820] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.651886] env[63175]: DEBUG nova.network.neutron [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.859377] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.941889] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa1e6fa-08b5-427c-8df6-c1a50fc5df13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.952021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a70aeb-a12f-4c72-a534-06d8d4a45d63 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.980918] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3e5175-e26b-4bf0-b09f-1bf0856c362b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.988612] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1183cac-03e7-4339-9017-801c83506f2f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.004540] env[63175]: DEBUG nova.compute.provider_tree [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.120894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Releasing lock "refresh_cache-16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.121152] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 609.121331] env[63175]: DEBUG nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 609.121497] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.151295] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.154362] env[63175]: DEBUG oslo_concurrency.lockutils [req-264e79ae-53f9-4468-b013-9b52e1dc3521 req-65d3a7fe-954b-4ca1-890e-3d7c97347530 service nova] Releasing lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.154784] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquired lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.154784] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.508969] env[63175]: DEBUG nova.scheduler.client.report [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 609.657318] env[63175]: DEBUG nova.network.neutron [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.677071] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.729813] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.017020] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.017020] env[63175]: ERROR nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Traceback (most recent call last): [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.driver.spawn(context, instance, image_meta, [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.017020] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] vm_ref = self.build_virtual_machine(instance, [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] for vif in network_info: [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self._sync_wrapper(fn, *args, **kwargs) [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.wait() [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self[:] = self._gt.wait() [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self._exit_event.wait() [ 610.017496] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] result = hub.switch() [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return self.greenlet.switch() [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] result = function(*args, **kwargs) [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] return func(*args, **kwargs) [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise e [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] nwinfo = self.network_api.allocate_for_instance( [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.017789] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] created_port_ids = self._update_ports_for_instance( [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] with excutils.save_and_reraise_exception(): [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] self.force_reraise() [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise self.value [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] updated_port = self._update_port( [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] _ensure_no_port_binding_failure(port) [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.018094] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] raise exception.PortBindingFailed(port_id=port['id']) [ 610.018352] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] nova.exception.PortBindingFailed: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. [ 610.018352] env[63175]: ERROR nova.compute.manager [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] [ 610.018352] env[63175]: DEBUG nova.compute.utils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.018863] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.089s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.022090] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Build of instance 7d227eb2-b270-4304-91ee-7ae20c33f633 was re-scheduled: Binding failed for port 64cc7a9c-4f5f-4e7a-9b87-b343b3fbdae2, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 610.022695] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 610.023087] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquiring lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.024495] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Acquired lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.024495] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.074229] env[63175]: DEBUG nova.compute.manager [req-74323023-983a-4921-aaaa-c7ea5b312e46 req-742c91a4-fab2-472a-b822-2e244cb45326 service nova] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Received event network-vif-deleted-f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 610.162410] env[63175]: INFO nova.compute.manager [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] [instance: 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf] Took 1.04 seconds to deallocate network for instance. [ 610.233289] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Releasing lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.233509] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 610.233737] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.234068] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77bf395b-839c-4da0-9b1d-44fafa9781aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.244266] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630555e0-c2c3-4004-8be4-8f65b840e40b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.266070] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8fa60fe5-a397-4b57-bb18-f41c0029743b could not be found. [ 610.266322] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.266502] env[63175]: INFO nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 610.266744] env[63175]: DEBUG oslo.service.loopingcall [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.266952] env[63175]: DEBUG nova.compute.manager [-] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 610.267054] env[63175]: DEBUG nova.network.neutron [-] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.293145] env[63175]: DEBUG nova.network.neutron [-] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.548269] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.721022] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.795381] env[63175]: DEBUG nova.network.neutron [-] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.944023] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5d20cd-45b3-4a71-9ce6-4b1b212e6d98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.951552] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce76b884-d31c-4ad3-8701-a160f49a46bd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.980318] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999f9e0c-c9c7-423f-a315-53635721cd5b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.987493] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36564472-97bc-4f46-82d8-7e29716b10d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.000270] env[63175]: DEBUG nova.compute.provider_tree [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.196079] env[63175]: INFO nova.scheduler.client.report [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Deleted allocations for instance 16e6ac19-b902-4db9-b5a5-5e351cbb3dcf [ 611.226057] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Releasing lock "refresh_cache-7d227eb2-b270-4304-91ee-7ae20c33f633" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.226318] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 611.226503] env[63175]: DEBUG nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 611.226669] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.276450] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.299882] env[63175]: INFO nova.compute.manager [-] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Took 1.03 seconds to deallocate network for instance. [ 611.302689] env[63175]: DEBUG nova.compute.claims [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.302959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.503541] env[63175]: DEBUG nova.scheduler.client.report [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 611.709894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d51522f-5359-4eeb-8675-009d3ff19c5b tempest-VolumesAssistedSnapshotsTest-1106373230 tempest-VolumesAssistedSnapshotsTest-1106373230-project-member] Lock "16e6ac19-b902-4db9-b5a5-5e351cbb3dcf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.750s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.778804] env[63175]: DEBUG nova.network.neutron [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.016055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.016055] env[63175]: ERROR nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Traceback (most recent call last): [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.driver.spawn(context, instance, image_meta, [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.016055] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] vm_ref = self.build_virtual_machine(instance, [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] for vif in network_info: [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self._sync_wrapper(fn, *args, **kwargs) [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.wait() [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self[:] = self._gt.wait() [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self._exit_event.wait() [ 612.016657] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] result = hub.switch() [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return self.greenlet.switch() [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] result = function(*args, **kwargs) [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] return func(*args, **kwargs) [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise e [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] nwinfo = self.network_api.allocate_for_instance( [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.016945] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] created_port_ids = self._update_ports_for_instance( [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] with excutils.save_and_reraise_exception(): [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] self.force_reraise() [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise self.value [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] updated_port = self._update_port( [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] _ensure_no_port_binding_failure(port) [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.017242] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] raise exception.PortBindingFailed(port_id=port['id']) [ 612.017490] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] nova.exception.PortBindingFailed: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. [ 612.017490] env[63175]: ERROR nova.compute.manager [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] [ 612.017490] env[63175]: DEBUG nova.compute.utils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.017490] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.508s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.022250] env[63175]: INFO nova.compute.claims [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.026401] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Build of instance 3c50a11c-cab7-461f-9550-575c4f51a4e5 was re-scheduled: Binding failed for port 7d1da76b-3228-4b1b-a259-50d23cc5c021, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 612.026797] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 612.027073] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquiring lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.027198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Acquired lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.027918] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.212895] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 612.283766] env[63175]: INFO nova.compute.manager [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] [instance: 7d227eb2-b270-4304-91ee-7ae20c33f633] Took 1.05 seconds to deallocate network for instance. [ 612.568878] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.746553] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.748726] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.254022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Releasing lock "refresh_cache-3c50a11c-cab7-461f-9550-575c4f51a4e5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.254022] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 613.254022] env[63175]: DEBUG nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.254022] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.327052] env[63175]: INFO nova.scheduler.client.report [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Deleted allocations for instance 7d227eb2-b270-4304-91ee-7ae20c33f633 [ 613.361958] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.578984] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7edf192-6376-40e6-8076-78de2486c936 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.588584] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d635ec0-9310-4122-8133-a00387a8e0c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.631270] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50703215-ae22-4293-a3ee-819ce9fb9e6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.642084] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733d8708-8aa2-4ffa-9809-255d0e2fcc0c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.658790] env[63175]: DEBUG nova.compute.provider_tree [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.842509] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f55fdcdb-85e9-45ac-a1d5-c899a96ea405 tempest-ServerAddressesTestJSON-1984200958 tempest-ServerAddressesTestJSON-1984200958-project-member] Lock "7d227eb2-b270-4304-91ee-7ae20c33f633" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.786s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.864173] env[63175]: DEBUG nova.network.neutron [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.161915] env[63175]: DEBUG nova.scheduler.client.report [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 614.346911] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 614.368448] env[63175]: INFO nova.compute.manager [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] [instance: 3c50a11c-cab7-461f-9550-575c4f51a4e5] Took 1.12 seconds to deallocate network for instance. [ 614.667572] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.651s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.668145] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.671448] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.962s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.874900] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.179696] env[63175]: DEBUG nova.compute.utils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.187894] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 615.188178] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.190382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquiring lock "501f803b-f4ea-4777-909e-7bb808628cc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.190627] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "501f803b-f4ea-4777-909e-7bb808628cc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.280512] env[63175]: DEBUG nova.policy [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42b746fba46e4896aff31d5a4412ff2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ee5fc1bc9c54d32972dc2642eed9e9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.411200] env[63175]: INFO nova.scheduler.client.report [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Deleted allocations for instance 3c50a11c-cab7-461f-9550-575c4f51a4e5 [ 615.687152] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.718679] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1413fc0d-11e2-4ec2-ad25-9049d8c7ec33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.726729] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b962adb0-bcf6-4631-aecd-d086b81391e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.761394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e75b0d-720d-4921-a7fe-4c46251d5e65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.769577] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bbfb1a-988d-45ea-91bf-71187f8b703c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.783715] env[63175]: DEBUG nova.compute.provider_tree [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.926465] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6deae108-7f3a-4cf5-92df-b9e205399840 tempest-ServerActionsTestOtherA-1665754182 tempest-ServerActionsTestOtherA-1665754182-project-member] Lock "3c50a11c-cab7-461f-9550-575c4f51a4e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.575s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.191688] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Successfully created port: 3f6020a9-2073-4800-802c-3eb3805e2419 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.286512] env[63175]: DEBUG nova.scheduler.client.report [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 616.429999] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 616.504218] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.504475] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.701424] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.742712] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.742957] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.743298] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.743389] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.744286] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.744286] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.744286] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.744286] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.746378] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.746378] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.746476] env[63175]: DEBUG nova.virt.hardware [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.748303] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2cd07c-fc69-4a98-9763-21e7e312f2f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.757355] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d58e5e2-967b-48b1-a404-a1c078b33a86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.791480] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.120s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.792505] env[63175]: ERROR nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Traceback (most recent call last): [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.driver.spawn(context, instance, image_meta, [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] vm_ref = self.build_virtual_machine(instance, [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.792505] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] for vif in network_info: [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self._sync_wrapper(fn, *args, **kwargs) [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.wait() [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self[:] = self._gt.wait() [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self._exit_event.wait() [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] result = hub.switch() [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.792802] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return self.greenlet.switch() [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] result = function(*args, **kwargs) [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] return func(*args, **kwargs) [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise e [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] nwinfo = self.network_api.allocate_for_instance( [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] created_port_ids = self._update_ports_for_instance( [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] with excutils.save_and_reraise_exception(): [ 616.793099] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] self.force_reraise() [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise self.value [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] updated_port = self._update_port( [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] _ensure_no_port_binding_failure(port) [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] raise exception.PortBindingFailed(port_id=port['id']) [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] nova.exception.PortBindingFailed: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. [ 616.793380] env[63175]: ERROR nova.compute.manager [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] [ 616.793626] env[63175]: DEBUG nova.compute.utils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.796383] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.008s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.800835] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Build of instance 652f4c1b-69b4-415d-ab14-bbf10043471d was re-scheduled: Binding failed for port b8256eb0-ef2d-49ca-9ed4-c54312c9505a, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 616.801580] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 616.801987] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquiring lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.802305] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Acquired lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.802593] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.960242] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.315932] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "12a2d788-c995-468b-b2a2-17dba8ad01db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.316175] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.362504] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.447383] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.632285] env[63175]: DEBUG nova.compute.manager [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Received event network-changed-3f6020a9-2073-4800-802c-3eb3805e2419 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 617.632285] env[63175]: DEBUG nova.compute.manager [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Refreshing instance network info cache due to event network-changed-3f6020a9-2073-4800-802c-3eb3805e2419. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 617.632285] env[63175]: DEBUG oslo_concurrency.lockutils [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] Acquiring lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.632285] env[63175]: DEBUG oslo_concurrency.lockutils [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] Acquired lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.632285] env[63175]: DEBUG nova.network.neutron [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Refreshing network info cache for port 3f6020a9-2073-4800-802c-3eb3805e2419 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.646249] env[63175]: ERROR nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 617.646249] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.646249] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.646249] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.646249] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.646249] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.646249] env[63175]: ERROR nova.compute.manager raise self.value [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.646249] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.646249] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.646249] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.646648] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.646648] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.646648] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 617.646648] env[63175]: ERROR nova.compute.manager [ 617.646648] env[63175]: Traceback (most recent call last): [ 617.646648] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.646648] env[63175]: listener.cb(fileno) [ 617.646648] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.646648] env[63175]: result = function(*args, **kwargs) [ 617.646648] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.646648] env[63175]: return func(*args, **kwargs) [ 617.646648] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.646648] env[63175]: raise e [ 617.646648] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.646648] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 617.646648] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.646648] env[63175]: created_port_ids = self._update_ports_for_instance( [ 617.646648] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.646648] env[63175]: with excutils.save_and_reraise_exception(): [ 617.646648] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.646648] env[63175]: self.force_reraise() [ 617.646648] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.646648] env[63175]: raise self.value [ 617.646648] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.646648] env[63175]: updated_port = self._update_port( [ 617.646648] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.646648] env[63175]: _ensure_no_port_binding_failure(port) [ 617.646648] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.646648] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.648652] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 617.648652] env[63175]: Removing descriptor: 17 [ 617.648652] env[63175]: ERROR nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Traceback (most recent call last): [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] yield resources [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.driver.spawn(context, instance, image_meta, [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.648652] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] vm_ref = self.build_virtual_machine(instance, [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] for vif in network_info: [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self._sync_wrapper(fn, *args, **kwargs) [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.wait() [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self[:] = self._gt.wait() [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self._exit_event.wait() [ 617.649105] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] result = hub.switch() [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self.greenlet.switch() [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] result = function(*args, **kwargs) [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return func(*args, **kwargs) [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise e [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] nwinfo = self.network_api.allocate_for_instance( [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.649561] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] created_port_ids = self._update_ports_for_instance( [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] with excutils.save_and_reraise_exception(): [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.force_reraise() [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise self.value [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] updated_port = self._update_port( [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] _ensure_no_port_binding_failure(port) [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.650072] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise exception.PortBindingFailed(port_id=port['id']) [ 617.650447] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 617.650447] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] [ 617.650447] env[63175]: INFO nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Terminating instance [ 617.936629] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e358697a-fbea-44f4-a7d5-483907554618 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.947712] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed7cf62-ca55-4314-afcd-927edd8d84ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.953237] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Releasing lock "refresh_cache-652f4c1b-69b4-415d-ab14-bbf10043471d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.953493] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 617.953681] env[63175]: DEBUG nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 617.953814] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.990087] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.992035] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056154b4-7b07-4a58-a963-1da6fb940b23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.002692] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdcafaf-2c0f-4bd2-bfe2-6d4109548e6c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.019021] env[63175]: DEBUG nova.compute.provider_tree [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.152253] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquiring lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.160253] env[63175]: DEBUG nova.network.neutron [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.284860] env[63175]: DEBUG nova.network.neutron [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.496773] env[63175]: DEBUG nova.network.neutron [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.522671] env[63175]: DEBUG nova.scheduler.client.report [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 618.792389] env[63175]: DEBUG oslo_concurrency.lockutils [req-6480a2a4-bc14-400b-b129-945a24cd3860 req-16cc2b74-989d-4602-8a48-4a976c5764e4 service nova] Releasing lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.793900] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquired lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.793900] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.000361] env[63175]: INFO nova.compute.manager [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] [instance: 652f4c1b-69b4-415d-ab14-bbf10043471d] Took 1.05 seconds to deallocate network for instance. [ 619.027576] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.231s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.028272] env[63175]: ERROR nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Traceback (most recent call last): [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.driver.spawn(context, instance, image_meta, [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] vm_ref = self.build_virtual_machine(instance, [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.028272] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] for vif in network_info: [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self._sync_wrapper(fn, *args, **kwargs) [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.wait() [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self[:] = self._gt.wait() [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self._exit_event.wait() [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] result = hub.switch() [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.028538] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return self.greenlet.switch() [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] result = function(*args, **kwargs) [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] return func(*args, **kwargs) [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise e [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] nwinfo = self.network_api.allocate_for_instance( [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] created_port_ids = self._update_ports_for_instance( [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] with excutils.save_and_reraise_exception(): [ 619.028903] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] self.force_reraise() [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise self.value [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] updated_port = self._update_port( [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] _ensure_no_port_binding_failure(port) [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] raise exception.PortBindingFailed(port_id=port['id']) [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] nova.exception.PortBindingFailed: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. [ 619.029245] env[63175]: ERROR nova.compute.manager [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] [ 619.029534] env[63175]: DEBUG nova.compute.utils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.030678] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.236s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.034594] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Build of instance 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5 was re-scheduled: Binding failed for port 051a669a-01ac-4efc-8711-1e76c6cac8a8, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 619.035117] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 619.035357] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.035506] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.035662] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.318107] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.465316] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.587552] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.723137] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.739853] env[63175]: DEBUG nova.compute.manager [req-8423d465-d5fb-43d7-a4d8-7f08534cb40e req-6a7e4667-af13-45f6-883b-0bc8a672d509 service nova] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Received event network-vif-deleted-3f6020a9-2073-4800-802c-3eb3805e2419 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 619.980770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Releasing lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.981197] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 619.981397] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.981692] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15a731b3-668c-4c26-8595-ba5670faf5d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.995021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820ecc44-435b-4353-a302-6fa803c7d1c6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.018848] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4eb6653-174b-46f5-91a1-224587243799 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.028088] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db596843-fecc-46b9-aba3-3926d93aa9ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.036871] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8 could not be found. [ 620.037273] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.038277] env[63175]: INFO nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Took 0.06 seconds to destroy the instance on the hypervisor. [ 620.038277] env[63175]: DEBUG oslo.service.loopingcall [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.038647] env[63175]: INFO nova.scheduler.client.report [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Deleted allocations for instance 652f4c1b-69b4-415d-ab14-bbf10043471d [ 620.047342] env[63175]: DEBUG nova.compute.manager [-] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 620.047452] env[63175]: DEBUG nova.network.neutron [-] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.080143] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a959e4-7e6a-46b4-a9ae-1c8ee72f510a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.089715] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13b13f5-c85e-48a7-85d7-b0c73fb9864f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.106198] env[63175]: DEBUG nova.compute.provider_tree [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.106675] env[63175]: DEBUG nova.network.neutron [-] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.226685] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.226933] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 620.227139] env[63175]: DEBUG nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 620.227317] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.245715] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.577968] env[63175]: DEBUG oslo_concurrency.lockutils [None req-beec83c1-c7d7-4810-9ee7-a03203c00fe4 tempest-ServerDiagnosticsNegativeTest-826916697 tempest-ServerDiagnosticsNegativeTest-826916697-project-member] Lock "652f4c1b-69b4-415d-ab14-bbf10043471d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.047s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.609502] env[63175]: DEBUG nova.scheduler.client.report [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 620.615144] env[63175]: DEBUG nova.network.neutron [-] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.748713] env[63175]: DEBUG nova.network.neutron [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.082486] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 621.120611] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.089s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.120611] env[63175]: ERROR nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Traceback (most recent call last): [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.driver.spawn(context, instance, image_meta, [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.120611] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] vm_ref = self.build_virtual_machine(instance, [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] for vif in network_info: [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self._sync_wrapper(fn, *args, **kwargs) [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.wait() [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self[:] = self._gt.wait() [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self._exit_event.wait() [ 621.120964] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] result = hub.switch() [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return self.greenlet.switch() [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] result = function(*args, **kwargs) [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] return func(*args, **kwargs) [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise e [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] nwinfo = self.network_api.allocate_for_instance( [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.121317] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] created_port_ids = self._update_ports_for_instance( [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] with excutils.save_and_reraise_exception(): [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] self.force_reraise() [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise self.value [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] updated_port = self._update_port( [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] _ensure_no_port_binding_failure(port) [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.121632] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] raise exception.PortBindingFailed(port_id=port['id']) [ 621.121944] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] nova.exception.PortBindingFailed: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. [ 621.121944] env[63175]: ERROR nova.compute.manager [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] [ 621.121944] env[63175]: DEBUG nova.compute.utils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.123398] env[63175]: INFO nova.compute.manager [-] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Took 1.07 seconds to deallocate network for instance. [ 621.123730] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.747s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.125184] env[63175]: INFO nova.compute.claims [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.130790] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Build of instance 4883dd7f-a13e-4efc-90ff-ca42909a7f0c was re-scheduled: Binding failed for port 71cf3fb8-b22b-4285-a520-b7a96bdd17bc, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 621.131255] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 621.131343] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.131438] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquired lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.132784] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.140551] env[63175]: DEBUG nova.compute.claims [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.140852] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.252816] env[63175]: INFO nova.compute.manager [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5] Took 1.02 seconds to deallocate network for instance. [ 621.604479] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.668508] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.777349] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.279062] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Releasing lock "refresh_cache-4883dd7f-a13e-4efc-90ff-ca42909a7f0c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.280713] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 622.280713] env[63175]: DEBUG nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 622.280713] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.298703] env[63175]: INFO nova.scheduler.client.report [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance 6d1c9a09-a5e6-4f2f-b79c-dde271b955f5 [ 622.306184] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.576123] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1267eaea-0635-43cf-a578-3548e7faacbd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.582664] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190b5483-c455-49fb-9db5-794186a8c447 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.612774] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d80a3a-2e6e-46bb-b84f-e47bb26d1ce1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.619789] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6f206e-7c4c-4952-8a91-1f32ceddbfe2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.632806] env[63175]: DEBUG nova.compute.provider_tree [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.793027] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.793308] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.811671] env[63175]: DEBUG nova.network.neutron [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.814629] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f45cf0f3-4f1a-4f51-83de-a27f2f301186 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "6d1c9a09-a5e6-4f2f-b79c-dde271b955f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.574s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.136158] env[63175]: DEBUG nova.scheduler.client.report [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 623.314338] env[63175]: INFO nova.compute.manager [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 4883dd7f-a13e-4efc-90ff-ca42909a7f0c] Took 1.03 seconds to deallocate network for instance. [ 623.318690] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 623.641448] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.642311] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 623.644845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.706s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.646225] env[63175]: INFO nova.compute.claims [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.838574] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.863206] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.863483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.150783] env[63175]: DEBUG nova.compute.utils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.154012] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 624.154195] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.212735] env[63175]: DEBUG nova.policy [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4088b59ec3034a679e928ca63d0f20f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52cea36517e548db8a19e0d9bd1e2381', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.339805] env[63175]: INFO nova.scheduler.client.report [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Deleted allocations for instance 4883dd7f-a13e-4efc-90ff-ca42909a7f0c [ 624.655632] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 624.698208] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Successfully created port: ff1f91f9-09c9-4d60-834b-5a6665f5d9b6 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.854377] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d962127d-f9aa-4f8c-af62-c38bd5c48816 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "4883dd7f-a13e-4efc-90ff-ca42909a7f0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.760s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.071906] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91679962-1b0d-411d-bb8b-5fbcca285f69 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.079784] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8291cde5-c1d9-4cf7-af18-fc870b645297 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.110321] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c915ec09-f9df-4e76-b176-1058d95f4af3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.117648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9dbda9-7a7d-4131-8d5d-d4536cd5119f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.131316] env[63175]: DEBUG nova.compute.provider_tree [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.355941] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 625.492859] env[63175]: DEBUG nova.compute.manager [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Received event network-changed-ff1f91f9-09c9-4d60-834b-5a6665f5d9b6 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 625.493175] env[63175]: DEBUG nova.compute.manager [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Refreshing instance network info cache due to event network-changed-ff1f91f9-09c9-4d60-834b-5a6665f5d9b6. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 625.493542] env[63175]: DEBUG oslo_concurrency.lockutils [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] Acquiring lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.493694] env[63175]: DEBUG oslo_concurrency.lockutils [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] Acquired lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.493903] env[63175]: DEBUG nova.network.neutron [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Refreshing network info cache for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.637123] env[63175]: DEBUG nova.scheduler.client.report [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 625.675380] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 625.703056] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.703056] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.703056] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.703319] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.703319] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.703319] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.703319] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.703319] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.704115] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.704907] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.704907] env[63175]: DEBUG nova.virt.hardware [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.705635] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe44f763-bcc5-4045-87df-b7797243a2c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.714161] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38286569-3faa-4c8e-bffa-e785a012c846 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.718819] env[63175]: ERROR nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 625.718819] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.718819] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.718819] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.718819] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.718819] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.718819] env[63175]: ERROR nova.compute.manager raise self.value [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.718819] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.718819] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.718819] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.719360] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.719360] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.719360] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 625.719360] env[63175]: ERROR nova.compute.manager [ 625.719360] env[63175]: Traceback (most recent call last): [ 625.719360] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.719360] env[63175]: listener.cb(fileno) [ 625.719360] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.719360] env[63175]: result = function(*args, **kwargs) [ 625.719360] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.719360] env[63175]: return func(*args, **kwargs) [ 625.719360] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.719360] env[63175]: raise e [ 625.719360] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.719360] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 625.719360] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.719360] env[63175]: created_port_ids = self._update_ports_for_instance( [ 625.719360] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.719360] env[63175]: with excutils.save_and_reraise_exception(): [ 625.719360] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.719360] env[63175]: self.force_reraise() [ 625.719360] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.719360] env[63175]: raise self.value [ 625.719360] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.719360] env[63175]: updated_port = self._update_port( [ 625.719360] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.719360] env[63175]: _ensure_no_port_binding_failure(port) [ 625.719360] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.719360] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.720644] env[63175]: nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 625.720644] env[63175]: Removing descriptor: 17 [ 625.730702] env[63175]: ERROR nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Traceback (most recent call last): [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] yield resources [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.driver.spawn(context, instance, image_meta, [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] vm_ref = self.build_virtual_machine(instance, [ 625.730702] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] for vif in network_info: [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return self._sync_wrapper(fn, *args, **kwargs) [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.wait() [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self[:] = self._gt.wait() [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return self._exit_event.wait() [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 625.731330] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] current.throw(*self._exc) [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] result = function(*args, **kwargs) [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return func(*args, **kwargs) [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise e [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] nwinfo = self.network_api.allocate_for_instance( [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] created_port_ids = self._update_ports_for_instance( [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] with excutils.save_and_reraise_exception(): [ 625.731787] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.force_reraise() [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise self.value [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] updated_port = self._update_port( [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] _ensure_no_port_binding_failure(port) [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise exception.PortBindingFailed(port_id=port['id']) [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 625.732201] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] [ 625.732201] env[63175]: INFO nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Terminating instance [ 625.880072] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.909202] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.909202] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.012122] env[63175]: DEBUG nova.network.neutron [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.084860] env[63175]: DEBUG nova.network.neutron [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.140239] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.141600] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 626.146035] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.840s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.235683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.415373] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.415578] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 626.415703] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 626.587209] env[63175]: DEBUG oslo_concurrency.lockutils [req-517fad9a-6e8c-440d-89bc-9dad1f1b0ca3 req-13e87332-c72d-4006-8413-154127e6e109 service nova] Releasing lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.587635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.587838] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 626.647742] env[63175]: DEBUG nova.compute.utils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.649550] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 626.649731] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 626.743830] env[63175]: DEBUG nova.policy [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afe7974a141b401a82f7f352825fddb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '171e1987d09940dcb67643a6b7bbf2b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.920795] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 626.921084] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 626.921134] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 626.922031] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 626.922031] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922031] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922031] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922031] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922257] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922257] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.922934] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 626.922934] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.083300] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Successfully created port: d052f968-ecf5-4a74-ad38-ac1267ea32f4 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.108346] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.125720] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf998cf-7afa-457e-976a-2817d2645a79 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.133797] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f862b757-6805-4d9b-b0c9-e0c97ef61433 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.168994] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 627.172593] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05df2c15-4978-4243-b68e-920088c65786 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.181736] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8add190-ed35-450c-8c4c-65c55a550000 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.196171] env[63175]: DEBUG nova.compute.provider_tree [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.259866] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.429877] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.523805] env[63175]: DEBUG nova.compute.manager [req-6461c5e6-e39f-4610-9bf4-cffc95cbc5e9 req-53c1340a-9fea-4c03-b10b-520e1b1564b9 service nova] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Received event network-vif-deleted-ff1f91f9-09c9-4d60-834b-5a6665f5d9b6 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 627.704030] env[63175]: DEBUG nova.scheduler.client.report [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 627.765203] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.768627] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 627.768627] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.768627] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d16b4067-26a1-440d-9264-61e4dc4bec16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.781583] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68fb88f-f1fd-4623-a1ce-69eba66faf62 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.804694] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1939e74b-3ab5-428b-86c5-4d6de694c454 could not be found. [ 627.805029] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.806163] env[63175]: INFO nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Took 0.04 seconds to destroy the instance on the hypervisor. [ 627.806163] env[63175]: DEBUG oslo.service.loopingcall [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.806163] env[63175]: DEBUG nova.compute.manager [-] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 627.806163] env[63175]: DEBUG nova.network.neutron [-] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.825668] env[63175]: DEBUG nova.network.neutron [-] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.112900] env[63175]: ERROR nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 628.112900] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.112900] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.112900] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.112900] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.112900] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.112900] env[63175]: ERROR nova.compute.manager raise self.value [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.112900] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.112900] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.112900] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.113947] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.113947] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.113947] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 628.113947] env[63175]: ERROR nova.compute.manager [ 628.113947] env[63175]: Traceback (most recent call last): [ 628.113947] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.113947] env[63175]: listener.cb(fileno) [ 628.113947] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.113947] env[63175]: result = function(*args, **kwargs) [ 628.113947] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.113947] env[63175]: return func(*args, **kwargs) [ 628.113947] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.113947] env[63175]: raise e [ 628.113947] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.113947] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 628.113947] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.113947] env[63175]: created_port_ids = self._update_ports_for_instance( [ 628.113947] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.113947] env[63175]: with excutils.save_and_reraise_exception(): [ 628.113947] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.113947] env[63175]: self.force_reraise() [ 628.113947] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.113947] env[63175]: raise self.value [ 628.113947] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.113947] env[63175]: updated_port = self._update_port( [ 628.113947] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.113947] env[63175]: _ensure_no_port_binding_failure(port) [ 628.113947] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.113947] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.115136] env[63175]: nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 628.115136] env[63175]: Removing descriptor: 17 [ 628.182378] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 628.207666] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.207950] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.208139] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.208336] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.208488] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.208637] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.208847] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.209012] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.209196] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.209393] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.209589] env[63175]: DEBUG nova.virt.hardware [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.210357] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.211248] env[63175]: ERROR nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Traceback (most recent call last): [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.driver.spawn(context, instance, image_meta, [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] vm_ref = self.build_virtual_machine(instance, [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.211248] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] for vif in network_info: [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self._sync_wrapper(fn, *args, **kwargs) [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.wait() [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self[:] = self._gt.wait() [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self._exit_event.wait() [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] result = hub.switch() [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.211606] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return self.greenlet.switch() [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] result = function(*args, **kwargs) [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] return func(*args, **kwargs) [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise e [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] nwinfo = self.network_api.allocate_for_instance( [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] created_port_ids = self._update_ports_for_instance( [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] with excutils.save_and_reraise_exception(): [ 628.211910] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] self.force_reraise() [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise self.value [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] updated_port = self._update_port( [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] _ensure_no_port_binding_failure(port) [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] raise exception.PortBindingFailed(port_id=port['id']) [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] nova.exception.PortBindingFailed: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. [ 628.212283] env[63175]: ERROR nova.compute.manager [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] [ 628.212547] env[63175]: DEBUG nova.compute.utils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.213580] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f61a68-cf54-4b32-8821-0ff86cd783a5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.216286] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.468s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.217964] env[63175]: INFO nova.compute.claims [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.227157] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d886bc35-877c-4836-8ce8-e59d996887d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.231297] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Build of instance 8fa60fe5-a397-4b57-bb18-f41c0029743b was re-scheduled: Binding failed for port f75ad3ef-a8b7-40ad-95fd-31c733bfb0fd, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 628.231764] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 628.231995] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquiring lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.232220] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Acquired lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.232314] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.244351] env[63175]: ERROR nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Traceback (most recent call last): [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] yield resources [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.driver.spawn(context, instance, image_meta, [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] vm_ref = self.build_virtual_machine(instance, [ 628.244351] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] for vif in network_info: [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return self._sync_wrapper(fn, *args, **kwargs) [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.wait() [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self[:] = self._gt.wait() [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return self._exit_event.wait() [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 628.244724] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] current.throw(*self._exc) [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] result = function(*args, **kwargs) [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return func(*args, **kwargs) [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise e [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] nwinfo = self.network_api.allocate_for_instance( [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] created_port_ids = self._update_ports_for_instance( [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] with excutils.save_and_reraise_exception(): [ 628.245103] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.force_reraise() [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise self.value [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] updated_port = self._update_port( [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] _ensure_no_port_binding_failure(port) [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise exception.PortBindingFailed(port_id=port['id']) [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 628.245584] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] [ 628.245584] env[63175]: INFO nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Terminating instance [ 628.327886] env[63175]: DEBUG nova.network.neutron [-] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.750963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.750963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquired lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.750963] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.752727] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.819999] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.830234] env[63175]: INFO nova.compute.manager [-] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Took 1.02 seconds to deallocate network for instance. [ 628.833581] env[63175]: DEBUG nova.compute.claims [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.833887] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.011982] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "67e8715d-b729-4013-8cca-44eaa55c662e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.012319] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.276054] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.323635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Releasing lock "refresh_cache-8fa60fe5-a397-4b57-bb18-f41c0029743b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.323961] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 629.324176] env[63175]: DEBUG nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 629.324345] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.340589] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.342301] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.590127] env[63175]: DEBUG nova.compute.manager [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Received event network-changed-d052f968-ecf5-4a74-ad38-ac1267ea32f4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 629.590318] env[63175]: DEBUG nova.compute.manager [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Refreshing instance network info cache due to event network-changed-d052f968-ecf5-4a74-ad38-ac1267ea32f4. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 629.590503] env[63175]: DEBUG oslo_concurrency.lockutils [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] Acquiring lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.681065] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760acc10-1739-469c-bee7-ed9beaf3fa04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.689241] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b175703f-6231-4fe2-bd0e-b00d08d24c67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.721119] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929b796d-6ec6-4584-ab9d-1e201ecad488 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.729840] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289e057c-3b4f-4687-8735-8c4a3aa7bbf4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.745640] env[63175]: DEBUG nova.compute.provider_tree [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.847915] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Releasing lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.848400] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 629.848599] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.848956] env[63175]: DEBUG nova.network.neutron [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.850187] env[63175]: DEBUG oslo_concurrency.lockutils [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] Acquired lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.850369] env[63175]: DEBUG nova.network.neutron [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Refreshing network info cache for port d052f968-ecf5-4a74-ad38-ac1267ea32f4 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 629.851296] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a6e5ae4-f238-40df-98a5-899d5845120b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.860466] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8901fcb9-8dac-4d6c-808e-7115e5684f1e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.882712] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6cd44d90-9092-483b-ab80-442e8f59435e could not be found. [ 629.882943] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.883147] env[63175]: INFO nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 629.883430] env[63175]: DEBUG oslo.service.loopingcall [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.883648] env[63175]: DEBUG nova.compute.manager [-] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 629.883756] env[63175]: DEBUG nova.network.neutron [-] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.899558] env[63175]: DEBUG nova.network.neutron [-] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.248669] env[63175]: DEBUG nova.scheduler.client.report [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 630.354617] env[63175]: INFO nova.compute.manager [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] [instance: 8fa60fe5-a397-4b57-bb18-f41c0029743b] Took 1.03 seconds to deallocate network for instance. [ 630.374447] env[63175]: DEBUG nova.network.neutron [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.401566] env[63175]: DEBUG nova.network.neutron [-] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.437602] env[63175]: DEBUG nova.network.neutron [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.753466] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.754063] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 630.756654] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.882s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.757976] env[63175]: INFO nova.compute.claims [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.904102] env[63175]: INFO nova.compute.manager [-] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Took 1.02 seconds to deallocate network for instance. [ 630.906216] env[63175]: DEBUG nova.compute.claims [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.906394] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.939849] env[63175]: DEBUG oslo_concurrency.lockutils [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] Releasing lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.940140] env[63175]: DEBUG nova.compute.manager [req-56e88d0d-e298-4490-9978-c5979221f38b req-320e2536-ac33-4925-a092-b3b531dd73e9 service nova] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Received event network-vif-deleted-d052f968-ecf5-4a74-ad38-ac1267ea32f4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 631.262848] env[63175]: DEBUG nova.compute.utils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.265946] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 631.266132] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.304226] env[63175]: DEBUG nova.policy [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4088b59ec3034a679e928ca63d0f20f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52cea36517e548db8a19e0d9bd1e2381', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.384134] env[63175]: INFO nova.scheduler.client.report [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Deleted allocations for instance 8fa60fe5-a397-4b57-bb18-f41c0029743b [ 631.585612] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Successfully created port: 9c863723-11d5-4577-8c1c-931b94e14451 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.771969] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 631.892499] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3a4beef5-c3dc-4c30-bf13-cc15a176fd27 tempest-ListImageFiltersTestJSON-599554689 tempest-ListImageFiltersTestJSON-599554689-project-member] Lock "8fa60fe5-a397-4b57-bb18-f41c0029743b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.395s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.210010] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc93e11f-a56b-4ab2-b5aa-2d7b299ef7a0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.217389] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3891bf5f-c64f-4dce-9725-0175127be6ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.247280] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb8b07c-74b8-42a8-a3e0-0a03578dbe6a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.254502] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7049a26c-d54a-489a-9df3-6ee15754ace6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.267636] env[63175]: DEBUG nova.compute.provider_tree [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.394773] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 632.477277] env[63175]: DEBUG nova.compute.manager [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Received event network-changed-9c863723-11d5-4577-8c1c-931b94e14451 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 632.477277] env[63175]: DEBUG nova.compute.manager [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Refreshing instance network info cache due to event network-changed-9c863723-11d5-4577-8c1c-931b94e14451. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 632.479842] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] Acquiring lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.480165] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] Acquired lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.480500] env[63175]: DEBUG nova.network.neutron [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Refreshing network info cache for port 9c863723-11d5-4577-8c1c-931b94e14451 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.662611] env[63175]: ERROR nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 632.662611] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.662611] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.662611] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.662611] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.662611] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.662611] env[63175]: ERROR nova.compute.manager raise self.value [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.662611] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.662611] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.662611] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.663122] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.663122] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.663122] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 632.663122] env[63175]: ERROR nova.compute.manager [ 632.663122] env[63175]: Traceback (most recent call last): [ 632.663122] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.663122] env[63175]: listener.cb(fileno) [ 632.663122] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.663122] env[63175]: result = function(*args, **kwargs) [ 632.663122] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.663122] env[63175]: return func(*args, **kwargs) [ 632.663122] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.663122] env[63175]: raise e [ 632.663122] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.663122] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 632.663122] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.663122] env[63175]: created_port_ids = self._update_ports_for_instance( [ 632.663122] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.663122] env[63175]: with excutils.save_and_reraise_exception(): [ 632.663122] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.663122] env[63175]: self.force_reraise() [ 632.663122] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.663122] env[63175]: raise self.value [ 632.663122] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.663122] env[63175]: updated_port = self._update_port( [ 632.663122] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.663122] env[63175]: _ensure_no_port_binding_failure(port) [ 632.663122] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.663122] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.663923] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 632.663923] env[63175]: Removing descriptor: 15 [ 632.773150] env[63175]: DEBUG nova.scheduler.client.report [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 632.780954] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 632.805588] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.805865] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.806108] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.806313] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.806461] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.806633] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.806840] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.807012] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.807173] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.807359] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.807533] env[63175]: DEBUG nova.virt.hardware [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.808543] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9151a71e-d585-4f24-9df4-5bb30081d677 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.817029] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba612a43-8543-40e3-8c61-683f91e3f0d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.833467] env[63175]: ERROR nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Traceback (most recent call last): [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] yield resources [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.driver.spawn(context, instance, image_meta, [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] vm_ref = self.build_virtual_machine(instance, [ 632.833467] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] for vif in network_info: [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return self._sync_wrapper(fn, *args, **kwargs) [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.wait() [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self[:] = self._gt.wait() [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return self._exit_event.wait() [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.833810] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] current.throw(*self._exc) [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] result = function(*args, **kwargs) [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return func(*args, **kwargs) [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise e [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] nwinfo = self.network_api.allocate_for_instance( [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] created_port_ids = self._update_ports_for_instance( [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] with excutils.save_and_reraise_exception(): [ 632.834189] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.force_reraise() [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise self.value [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] updated_port = self._update_port( [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] _ensure_no_port_binding_failure(port) [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise exception.PortBindingFailed(port_id=port['id']) [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 632.834525] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] [ 632.834525] env[63175]: INFO nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Terminating instance [ 632.918846] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.002796] env[63175]: DEBUG nova.network.neutron [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.101424] env[63175]: DEBUG nova.network.neutron [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.278856] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.279465] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 633.282492] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.322s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.286189] env[63175]: INFO nova.compute.claims [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.338816] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.605522] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9f68e9-1ef1-46a3-8b27-613015b25194 req-d9b6ca42-e018-406d-8712-f19e6ba82929 service nova] Releasing lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.605925] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.606124] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.791343] env[63175]: DEBUG nova.compute.utils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.793292] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 633.793584] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.865691] env[63175]: DEBUG nova.policy [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '484710bba41146459d7ae0ab92827817', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fcd017c8a444c3891f2eb4fee2452f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.125535] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.169696] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Successfully created port: 51cb423f-8eb2-47da-8647-453018aef00c {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.223729] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.298289] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 634.604460] env[63175]: DEBUG nova.compute.manager [req-8c078fbd-1b67-4878-a166-55041cf032c4 req-26ff52cf-a0ea-49e6-9bed-ac1091ca0a68 service nova] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Received event network-vif-deleted-9c863723-11d5-4577-8c1c-931b94e14451 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 634.724970] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f11ef7-6d57-4ee8-a593-fa470726f5f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.727844] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.728266] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 634.728457] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.728705] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf110189-c4ce-4042-b34a-c2ac36b8b106 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.736404] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5a7705-0d5b-4270-9534-0514259a7bc0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.741882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c69e61-152d-4943-94c3-915ae4f172cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.782235] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5805e4-dcfa-4a22-b5d2-8ffadc96c3bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.786136] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ce379bc8-add6-4008-b4db-c8bdb61a8e97 could not be found. [ 634.786136] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.786136] env[63175]: INFO nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Took 0.06 seconds to destroy the instance on the hypervisor. [ 634.786136] env[63175]: DEBUG oslo.service.loopingcall [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.786136] env[63175]: DEBUG nova.compute.manager [-] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.786343] env[63175]: DEBUG nova.network.neutron [-] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.792469] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423f4454-206d-4e77-9733-c3b178c753b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.809025] env[63175]: DEBUG nova.compute.provider_tree [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.810341] env[63175]: DEBUG nova.network.neutron [-] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.095564] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "36089589-d105-49e0-8ae7-790c814b036c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.095779] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.140181] env[63175]: ERROR nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 635.140181] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.140181] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.140181] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.140181] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.140181] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.140181] env[63175]: ERROR nova.compute.manager raise self.value [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.140181] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.140181] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.140181] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.140659] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.140659] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.140659] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 635.140659] env[63175]: ERROR nova.compute.manager [ 635.140659] env[63175]: Traceback (most recent call last): [ 635.140659] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.140659] env[63175]: listener.cb(fileno) [ 635.140659] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.140659] env[63175]: result = function(*args, **kwargs) [ 635.140659] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.140659] env[63175]: return func(*args, **kwargs) [ 635.140659] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.140659] env[63175]: raise e [ 635.140659] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.140659] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 635.140659] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.140659] env[63175]: created_port_ids = self._update_ports_for_instance( [ 635.140659] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.140659] env[63175]: with excutils.save_and_reraise_exception(): [ 635.140659] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.140659] env[63175]: self.force_reraise() [ 635.140659] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.140659] env[63175]: raise self.value [ 635.140659] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.140659] env[63175]: updated_port = self._update_port( [ 635.140659] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.140659] env[63175]: _ensure_no_port_binding_failure(port) [ 635.140659] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.140659] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.141472] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 635.141472] env[63175]: Removing descriptor: 15 [ 635.313807] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 635.318238] env[63175]: DEBUG nova.scheduler.client.report [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 635.321439] env[63175]: DEBUG nova.network.neutron [-] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.343010] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.343267] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.343422] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.343602] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.343746] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.343891] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.344110] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.344331] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.344521] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.344684] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.344851] env[63175]: DEBUG nova.virt.hardware [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.345893] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7730fb53-8f19-4bbf-b134-601711413647 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.354596] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcbe35a-9796-41cb-afdd-9af374296a4f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.368333] env[63175]: ERROR nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Traceback (most recent call last): [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] yield resources [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.driver.spawn(context, instance, image_meta, [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] vm_ref = self.build_virtual_machine(instance, [ 635.368333] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] for vif in network_info: [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return self._sync_wrapper(fn, *args, **kwargs) [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.wait() [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self[:] = self._gt.wait() [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return self._exit_event.wait() [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 635.368770] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] current.throw(*self._exc) [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] result = function(*args, **kwargs) [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return func(*args, **kwargs) [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise e [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] nwinfo = self.network_api.allocate_for_instance( [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] created_port_ids = self._update_ports_for_instance( [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] with excutils.save_and_reraise_exception(): [ 635.369133] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.force_reraise() [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise self.value [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] updated_port = self._update_port( [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] _ensure_no_port_binding_failure(port) [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise exception.PortBindingFailed(port_id=port['id']) [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 635.369479] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] [ 635.369479] env[63175]: INFO nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Terminating instance [ 635.824440] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.825656] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 635.827726] env[63175]: INFO nova.compute.manager [-] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Took 1.04 seconds to deallocate network for instance. [ 635.828263] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.687s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.832646] env[63175]: DEBUG nova.compute.claims [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.832870] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.877677] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquiring lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.877912] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquired lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.878127] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.330460] env[63175]: DEBUG nova.compute.utils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.331837] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 636.332012] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.383642] env[63175]: DEBUG nova.policy [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4088b59ec3034a679e928ca63d0f20f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52cea36517e548db8a19e0d9bd1e2381', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.399163] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.483832] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.640084] env[63175]: DEBUG nova.compute.manager [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Received event network-changed-51cb423f-8eb2-47da-8647-453018aef00c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 636.640296] env[63175]: DEBUG nova.compute.manager [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Refreshing instance network info cache due to event network-changed-51cb423f-8eb2-47da-8647-453018aef00c. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 636.640486] env[63175]: DEBUG oslo_concurrency.lockutils [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] Acquiring lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.664769] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Successfully created port: f42a568c-7569-41be-8170-db96e586cf1e {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.795866] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2245a64d-a942-4395-975c-d1a6adb8c329 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.803207] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436360ac-241b-4b23-a7d9-2ac45259aa76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.833145] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8ae105-0486-46bb-a98a-ba4c2acb9bd4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.835599] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 636.842484] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d34f872-ad28-4df8-be67-f6bf45193982 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.854994] env[63175]: DEBUG nova.compute.provider_tree [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.988428] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Releasing lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.988862] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 636.989078] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.989392] env[63175]: DEBUG oslo_concurrency.lockutils [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] Acquired lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.989563] env[63175]: DEBUG nova.network.neutron [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Refreshing network info cache for port 51cb423f-8eb2-47da-8647-453018aef00c {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.990707] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7524be44-d8f2-4592-8789-278209023afe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.999406] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eace818-970e-4c0a-b9a6-c8f154309d90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.021706] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 425a1271-cda7-45e4-89cb-cccd6231b623 could not be found. [ 637.021960] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.022181] env[63175]: INFO nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Took 0.03 seconds to destroy the instance on the hypervisor. [ 637.022448] env[63175]: DEBUG oslo.service.loopingcall [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.022672] env[63175]: DEBUG nova.compute.manager [-] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 637.022762] env[63175]: DEBUG nova.network.neutron [-] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.042370] env[63175]: DEBUG nova.network.neutron [-] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.357662] env[63175]: DEBUG nova.scheduler.client.report [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 637.507504] env[63175]: DEBUG nova.network.neutron [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.544114] env[63175]: DEBUG nova.network.neutron [-] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.580392] env[63175]: DEBUG nova.network.neutron [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.682569] env[63175]: ERROR nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 637.682569] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.682569] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.682569] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.682569] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.682569] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.682569] env[63175]: ERROR nova.compute.manager raise self.value [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.682569] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.682569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.682569] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.682993] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.682993] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.682993] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 637.682993] env[63175]: ERROR nova.compute.manager [ 637.682993] env[63175]: Traceback (most recent call last): [ 637.682993] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.682993] env[63175]: listener.cb(fileno) [ 637.682993] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.682993] env[63175]: result = function(*args, **kwargs) [ 637.682993] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.682993] env[63175]: return func(*args, **kwargs) [ 637.682993] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.682993] env[63175]: raise e [ 637.682993] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.682993] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 637.682993] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.682993] env[63175]: created_port_ids = self._update_ports_for_instance( [ 637.682993] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.682993] env[63175]: with excutils.save_and_reraise_exception(): [ 637.682993] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.682993] env[63175]: self.force_reraise() [ 637.682993] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.682993] env[63175]: raise self.value [ 637.682993] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.682993] env[63175]: updated_port = self._update_port( [ 637.682993] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.682993] env[63175]: _ensure_no_port_binding_failure(port) [ 637.682993] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.682993] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.683732] env[63175]: nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 637.683732] env[63175]: Removing descriptor: 15 [ 637.844740] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 637.862973] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.863666] env[63175]: ERROR nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Traceback (most recent call last): [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.driver.spawn(context, instance, image_meta, [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] vm_ref = self.build_virtual_machine(instance, [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.863666] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] for vif in network_info: [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self._sync_wrapper(fn, *args, **kwargs) [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.wait() [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self[:] = self._gt.wait() [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self._exit_event.wait() [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] result = hub.switch() [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.863978] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return self.greenlet.switch() [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] result = function(*args, **kwargs) [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] return func(*args, **kwargs) [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise e [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] nwinfo = self.network_api.allocate_for_instance( [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] created_port_ids = self._update_ports_for_instance( [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] with excutils.save_and_reraise_exception(): [ 637.864343] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] self.force_reraise() [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise self.value [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] updated_port = self._update_port( [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] _ensure_no_port_binding_failure(port) [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] raise exception.PortBindingFailed(port_id=port['id']) [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] nova.exception.PortBindingFailed: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. [ 637.864703] env[63175]: ERROR nova.compute.manager [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] [ 637.865013] env[63175]: DEBUG nova.compute.utils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.865647] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.261s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.869357] env[63175]: INFO nova.compute.claims [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.872903] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Build of instance dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8 was re-scheduled: Binding failed for port 3f6020a9-2073-4800-802c-3eb3805e2419, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 637.873431] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 637.873682] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquiring lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.873853] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Acquired lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.874066] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.877552] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.877821] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.878145] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.878380] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.878537] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.878687] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.878893] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.879092] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.879289] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.879450] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.879646] env[63175]: DEBUG nova.virt.hardware [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.880827] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cdff7e-d598-4e5a-8bc1-25a285cfe50d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.889360] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f75dbad-a2ca-4f49-a34e-472fecb4ccfd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.903243] env[63175]: ERROR nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Traceback (most recent call last): [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] yield resources [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.driver.spawn(context, instance, image_meta, [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] vm_ref = self.build_virtual_machine(instance, [ 637.903243] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] for vif in network_info: [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return self._sync_wrapper(fn, *args, **kwargs) [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.wait() [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self[:] = self._gt.wait() [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return self._exit_event.wait() [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 637.903626] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] current.throw(*self._exc) [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] result = function(*args, **kwargs) [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return func(*args, **kwargs) [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise e [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] nwinfo = self.network_api.allocate_for_instance( [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] created_port_ids = self._update_ports_for_instance( [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] with excutils.save_and_reraise_exception(): [ 637.904014] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.force_reraise() [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise self.value [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] updated_port = self._update_port( [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] _ensure_no_port_binding_failure(port) [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise exception.PortBindingFailed(port_id=port['id']) [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 637.904354] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] [ 637.904354] env[63175]: INFO nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Terminating instance [ 638.048880] env[63175]: INFO nova.compute.manager [-] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Took 1.03 seconds to deallocate network for instance. [ 638.051450] env[63175]: DEBUG nova.compute.claims [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.051672] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.082530] env[63175]: DEBUG oslo_concurrency.lockutils [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] Releasing lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.082773] env[63175]: DEBUG nova.compute.manager [req-2e49b451-68cc-441d-a14a-a27e713ec9f5 req-5776448d-ccfc-418a-b078-1ff012f266a4 service nova] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Received event network-vif-deleted-51cb423f-8eb2-47da-8647-453018aef00c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 638.391923] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.408050] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.408287] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.408479] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.488992] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.663238] env[63175]: DEBUG nova.compute.manager [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Received event network-changed-f42a568c-7569-41be-8170-db96e586cf1e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 638.663434] env[63175]: DEBUG nova.compute.manager [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Refreshing instance network info cache due to event network-changed-f42a568c-7569-41be-8170-db96e586cf1e. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 638.663620] env[63175]: DEBUG oslo_concurrency.lockutils [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] Acquiring lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.926126] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.991637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Releasing lock "refresh_cache-dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.991874] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 638.992094] env[63175]: DEBUG nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 638.992267] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.996343] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.007182] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.228182] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5e0595-f711-4ef3-b3f2-7689cafce440 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.235821] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fc23f1-8242-4c65-ae1a-e1281a5bf460 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.265741] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f721cb-b786-4769-8ee8-996f784e4fb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.272893] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c8d45b-992b-44a7-b201-d2bfa1c0d224 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.286740] env[63175]: DEBUG nova.compute.provider_tree [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.499087] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.499551] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 639.499751] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.501033] env[63175]: DEBUG oslo_concurrency.lockutils [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] Acquired lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.501033] env[63175]: DEBUG nova.network.neutron [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Refreshing network info cache for port f42a568c-7569-41be-8170-db96e586cf1e {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.501696] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ae61da1-f71c-4bcc-9875-8ea51cdc6bb3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.511306] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a86c98-3220-472a-9b79-465f4cf2fbba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.521224] env[63175]: DEBUG nova.network.neutron [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.533326] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5da92241-84f0-4510-b19d-b28cb57079ab could not be found. [ 639.534310] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.534310] env[63175]: INFO nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.534310] env[63175]: DEBUG oslo.service.loopingcall [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.534621] env[63175]: DEBUG nova.compute.manager [-] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 639.534719] env[63175]: DEBUG nova.network.neutron [-] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.549257] env[63175]: DEBUG nova.network.neutron [-] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.789546] env[63175]: DEBUG nova.scheduler.client.report [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 640.019599] env[63175]: DEBUG nova.network.neutron [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.023882] env[63175]: INFO nova.compute.manager [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] [instance: dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8] Took 1.03 seconds to deallocate network for instance. [ 640.052473] env[63175]: DEBUG nova.network.neutron [-] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.104139] env[63175]: DEBUG nova.network.neutron [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.295018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.295487] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 640.298173] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.460s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.300009] env[63175]: INFO nova.compute.claims [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.558059] env[63175]: INFO nova.compute.manager [-] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Took 1.02 seconds to deallocate network for instance. [ 640.560078] env[63175]: DEBUG nova.compute.claims [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.560274] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.606106] env[63175]: DEBUG oslo_concurrency.lockutils [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] Releasing lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.606433] env[63175]: DEBUG nova.compute.manager [req-46d3219c-9e1b-48a0-b3ec-24084ff6514a req-abeaa103-1962-4871-b286-2f0dd3ef29c6 service nova] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Received event network-vif-deleted-f42a568c-7569-41be-8170-db96e586cf1e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 640.804114] env[63175]: DEBUG nova.compute.utils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.807534] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 640.807748] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.860145] env[63175]: DEBUG nova.policy [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c654f31fd4bc49dc8e985bd75a811ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74dd39e20aad4ddb874282973cea02bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.053031] env[63175]: INFO nova.scheduler.client.report [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Deleted allocations for instance dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8 [ 641.167960] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Successfully created port: e988c082-794b-4c1d-b983-e69778caf0a4 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.308634] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 641.563084] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f3da00c9-d1a3-457c-afd1-59e841a44c26 tempest-ServersTestManualDisk-961404927 tempest-ServersTestManualDisk-961404927-project-member] Lock "dcea6ad2-4462-45d0-b66e-5b0dbbc02fa8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.326s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.766565] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de270ebf-c64b-4ca2-a897-3e52ff20a15c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.776430] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac8b6e0-e01c-4348-9353-c429dcb956c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.809599] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5956636-8e8c-4743-9f92-2fc96e945653 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.821228] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb424da-57be-4de4-a147-d5f6593f8d8f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.840760] env[63175]: DEBUG nova.compute.provider_tree [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.065274] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 642.099020] env[63175]: DEBUG nova.compute.manager [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Received event network-changed-e988c082-794b-4c1d-b983-e69778caf0a4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 642.099020] env[63175]: DEBUG nova.compute.manager [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Refreshing instance network info cache due to event network-changed-e988c082-794b-4c1d-b983-e69778caf0a4. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 642.099020] env[63175]: DEBUG oslo_concurrency.lockutils [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] Acquiring lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.099020] env[63175]: DEBUG oslo_concurrency.lockutils [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] Acquired lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.099020] env[63175]: DEBUG nova.network.neutron [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Refreshing network info cache for port e988c082-794b-4c1d-b983-e69778caf0a4 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 642.258030] env[63175]: ERROR nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 642.258030] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 642.258030] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.258030] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.258030] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.258030] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.258030] env[63175]: ERROR nova.compute.manager raise self.value [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.258030] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.258030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.258030] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.258477] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.258477] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.258477] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 642.258477] env[63175]: ERROR nova.compute.manager [ 642.258477] env[63175]: Traceback (most recent call last): [ 642.258477] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.258477] env[63175]: listener.cb(fileno) [ 642.258477] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.258477] env[63175]: result = function(*args, **kwargs) [ 642.258477] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.258477] env[63175]: return func(*args, **kwargs) [ 642.258477] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 642.258477] env[63175]: raise e [ 642.258477] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 642.258477] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 642.258477] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.258477] env[63175]: created_port_ids = self._update_ports_for_instance( [ 642.258477] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.258477] env[63175]: with excutils.save_and_reraise_exception(): [ 642.258477] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.258477] env[63175]: self.force_reraise() [ 642.258477] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.258477] env[63175]: raise self.value [ 642.258477] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.258477] env[63175]: updated_port = self._update_port( [ 642.258477] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.258477] env[63175]: _ensure_no_port_binding_failure(port) [ 642.258477] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.258477] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.259150] env[63175]: nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 642.259150] env[63175]: Removing descriptor: 17 [ 642.329020] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 642.343416] env[63175]: DEBUG nova.scheduler.client.report [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 642.359709] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.359984] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.360155] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.360355] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.360522] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.360642] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.360845] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.361040] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.361256] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.361388] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.361555] env[63175]: DEBUG nova.virt.hardware [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.362843] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880c0e5c-ab93-4df0-8707-ae6a6bf52ffe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.371556] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e294931f-b5a2-49d5-9df7-fe3d850a7f46 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.388104] env[63175]: ERROR nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Traceback (most recent call last): [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] yield resources [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.driver.spawn(context, instance, image_meta, [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] vm_ref = self.build_virtual_machine(instance, [ 642.388104] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] for vif in network_info: [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return self._sync_wrapper(fn, *args, **kwargs) [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.wait() [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self[:] = self._gt.wait() [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return self._exit_event.wait() [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.388692] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] current.throw(*self._exc) [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] result = function(*args, **kwargs) [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return func(*args, **kwargs) [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise e [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] nwinfo = self.network_api.allocate_for_instance( [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] created_port_ids = self._update_ports_for_instance( [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] with excutils.save_and_reraise_exception(): [ 642.389364] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.force_reraise() [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise self.value [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] updated_port = self._update_port( [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] _ensure_no_port_binding_failure(port) [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise exception.PortBindingFailed(port_id=port['id']) [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 642.389786] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] [ 642.389786] env[63175]: INFO nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Terminating instance [ 642.589551] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.619417] env[63175]: DEBUG nova.network.neutron [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.739702] env[63175]: DEBUG nova.network.neutron [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.856602] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.856602] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 642.857993] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.978s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.860142] env[63175]: INFO nova.compute.claims [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.893249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.243802] env[63175]: DEBUG oslo_concurrency.lockutils [req-32e33fef-eb50-44fb-8ded-60e46d6b86a2 req-d36d77a1-c9fc-4997-9fd3-b57191cca23b service nova] Releasing lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.244337] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.244762] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.366044] env[63175]: DEBUG nova.compute.utils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.369388] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 643.369686] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.425375] env[63175]: DEBUG nova.policy [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a3cc30cb1c94f00b66e5e5321e71cae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a13977fe862d4b58879c912fa7b7828b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.763090] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Successfully created port: fb241359-1ae8-4e7f-93ad-099af3f20e36 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.765519] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.865087] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.872610] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.125382] env[63175]: DEBUG nova.compute.manager [req-2fce9618-5bd6-4c49-a6da-4643c5adbfd0 req-7ef2d4e7-8bed-4761-bd76-103885a72821 service nova] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Received event network-vif-deleted-e988c082-794b-4c1d-b983-e69778caf0a4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 644.282968] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9e0ce0-2494-468f-aa2b-b717acae2a2e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.291629] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726bfebe-99aa-44b2-ab21-4a31b582bcf1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.321948] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99fa2e7-57e0-48eb-9b0c-789418ed77fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.329389] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f241a2da-6f9a-43b3-8daa-284a882ef12b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.342837] env[63175]: DEBUG nova.compute.provider_tree [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.366701] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.367190] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 644.367379] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.367643] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddab7e23-0ee2-49ee-8a20-e1597fe33a95 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.376227] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1334b9-8d4c-43a5-985b-06e4b61189e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.400751] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa could not be found. [ 644.400977] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.401172] env[63175]: INFO nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Took 0.03 seconds to destroy the instance on the hypervisor. [ 644.401450] env[63175]: DEBUG oslo.service.loopingcall [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.402022] env[63175]: DEBUG nova.compute.manager [-] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 644.402022] env[63175]: DEBUG nova.network.neutron [-] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.422506] env[63175]: DEBUG nova.network.neutron [-] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.845523] env[63175]: DEBUG nova.scheduler.client.report [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 644.892304] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 644.926522] env[63175]: DEBUG nova.network.neutron [-] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.935399] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.935399] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.935399] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.935680] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.935680] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.935680] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.935680] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.935680] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.935904] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.935904] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.938288] env[63175]: DEBUG nova.virt.hardware [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.939222] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de7d6f0-74f0-4c00-b7ca-84abad1e278c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.947820] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf7b846-5751-4f1e-80e3-9c0b8f85934d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.061655] env[63175]: ERROR nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 645.061655] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.061655] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.061655] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.061655] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.061655] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.061655] env[63175]: ERROR nova.compute.manager raise self.value [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.061655] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.061655] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.061655] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.062162] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.062162] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.062162] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 645.062162] env[63175]: ERROR nova.compute.manager [ 645.062162] env[63175]: Traceback (most recent call last): [ 645.062162] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.062162] env[63175]: listener.cb(fileno) [ 645.062162] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.062162] env[63175]: result = function(*args, **kwargs) [ 645.062162] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.062162] env[63175]: return func(*args, **kwargs) [ 645.062162] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.062162] env[63175]: raise e [ 645.062162] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.062162] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 645.062162] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.062162] env[63175]: created_port_ids = self._update_ports_for_instance( [ 645.062162] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.062162] env[63175]: with excutils.save_and_reraise_exception(): [ 645.062162] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.062162] env[63175]: self.force_reraise() [ 645.062162] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.062162] env[63175]: raise self.value [ 645.062162] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.062162] env[63175]: updated_port = self._update_port( [ 645.062162] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.062162] env[63175]: _ensure_no_port_binding_failure(port) [ 645.062162] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.062162] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.062866] env[63175]: nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 645.062866] env[63175]: Removing descriptor: 17 [ 645.063091] env[63175]: ERROR nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Traceback (most recent call last): [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] yield resources [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.driver.spawn(context, instance, image_meta, [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] vm_ref = self.build_virtual_machine(instance, [ 645.063091] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] for vif in network_info: [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self._sync_wrapper(fn, *args, **kwargs) [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.wait() [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self[:] = self._gt.wait() [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self._exit_event.wait() [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.063805] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] result = hub.switch() [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self.greenlet.switch() [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] result = function(*args, **kwargs) [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return func(*args, **kwargs) [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise e [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] nwinfo = self.network_api.allocate_for_instance( [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] created_port_ids = self._update_ports_for_instance( [ 645.064562] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] with excutils.save_and_reraise_exception(): [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.force_reraise() [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise self.value [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] updated_port = self._update_port( [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] _ensure_no_port_binding_failure(port) [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise exception.PortBindingFailed(port_id=port['id']) [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 645.065190] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] [ 645.070238] env[63175]: INFO nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Terminating instance [ 645.350227] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.350769] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 645.353357] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.924s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.356559] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.003s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.356781] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 645.357099] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.523s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.361074] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976aba17-4e72-43ec-b8fa-5a2aabcc996e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.369039] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ef8127-3b03-45ab-9a0f-ca82b108af07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.383929] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654c62c1-d81f-40a4-9d2d-1266d75d8d9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.390625] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83684afb-7d1b-4252-9ce0-d0ae07d3ffd3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.421606] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181528MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 645.421606] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.430443] env[63175]: INFO nova.compute.manager [-] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Took 1.03 seconds to deallocate network for instance. [ 645.434339] env[63175]: DEBUG nova.compute.claims [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 645.434529] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.568206] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquiring lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.568461] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquired lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.568643] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.862174] env[63175]: DEBUG nova.compute.utils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.863557] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 645.863727] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.911340] env[63175]: DEBUG nova.policy [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4cfacbbad794d2cb79f0fa821300bb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a05d0f203140474fb85ff32a5430332f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.092162] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.174143] env[63175]: DEBUG nova.compute.manager [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Received event network-changed-fb241359-1ae8-4e7f-93ad-099af3f20e36 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 646.174143] env[63175]: DEBUG nova.compute.manager [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Refreshing instance network info cache due to event network-changed-fb241359-1ae8-4e7f-93ad-099af3f20e36. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 646.174999] env[63175]: DEBUG oslo_concurrency.lockutils [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] Acquiring lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.214141] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Successfully created port: 6c88289f-bce6-4bda-a058-f6913ebf1a02 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.242385] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.300976] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f12f2ee-d854-423a-b2e1-56fd9fb4eb51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.311593] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5977adf6-1858-41d4-9144-26e956de9594 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.342203] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27475d6-9037-4827-b2e4-df17a470f0b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.352134] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b6d41b-61bb-44eb-939d-6223547d89cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.365351] env[63175]: DEBUG nova.compute.provider_tree [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.367326] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 646.750793] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Releasing lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.751166] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 646.751372] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.751711] env[63175]: DEBUG oslo_concurrency.lockutils [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] Acquired lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.751852] env[63175]: DEBUG nova.network.neutron [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Refreshing network info cache for port fb241359-1ae8-4e7f-93ad-099af3f20e36 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.752961] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-072ca2fe-2e1b-404e-b72d-5d944c9ef1b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.763486] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f1b6e3-6fdd-4cc7-bd2b-d80dc950d702 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.799659] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2cb7696b-f146-4db6-9e83-385fb8af1127 could not be found. [ 646.799659] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.799659] env[63175]: INFO nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Took 0.04 seconds to destroy the instance on the hypervisor. [ 646.799659] env[63175]: DEBUG oslo.service.loopingcall [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.799659] env[63175]: DEBUG nova.compute.manager [-] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 646.799659] env[63175]: DEBUG nova.network.neutron [-] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.836785] env[63175]: DEBUG nova.network.neutron [-] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.871341] env[63175]: DEBUG nova.scheduler.client.report [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 647.279223] env[63175]: DEBUG nova.network.neutron [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.349099] env[63175]: DEBUG nova.network.neutron [-] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.362847] env[63175]: DEBUG nova.network.neutron [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.382123] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.382728] env[63175]: ERROR nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Traceback (most recent call last): [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.driver.spawn(context, instance, image_meta, [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] vm_ref = self.build_virtual_machine(instance, [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.382728] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] for vif in network_info: [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return self._sync_wrapper(fn, *args, **kwargs) [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.wait() [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self[:] = self._gt.wait() [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return self._exit_event.wait() [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] current.throw(*self._exc) [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.383058] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] result = function(*args, **kwargs) [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] return func(*args, **kwargs) [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise e [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] nwinfo = self.network_api.allocate_for_instance( [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] created_port_ids = self._update_ports_for_instance( [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] with excutils.save_and_reraise_exception(): [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] self.force_reraise() [ 647.383368] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise self.value [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] updated_port = self._update_port( [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] _ensure_no_port_binding_failure(port) [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] raise exception.PortBindingFailed(port_id=port['id']) [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] nova.exception.PortBindingFailed: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. [ 647.383672] env[63175]: ERROR nova.compute.manager [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] [ 647.383672] env[63175]: DEBUG nova.compute.utils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 647.385251] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 647.387727] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.481s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.391896] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Build of instance 1939e74b-3ab5-428b-86c5-4d6de694c454 was re-scheduled: Binding failed for port ff1f91f9-09c9-4d60-834b-5a6665f5d9b6, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 647.395696] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 647.395696] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.395696] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.395696] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.408569] env[63175]: ERROR nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 647.408569] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.408569] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.408569] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.408569] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.408569] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.408569] env[63175]: ERROR nova.compute.manager raise self.value [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.408569] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.408569] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.408569] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.409039] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.409039] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.409039] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 647.409039] env[63175]: ERROR nova.compute.manager [ 647.409039] env[63175]: Traceback (most recent call last): [ 647.409039] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.409039] env[63175]: listener.cb(fileno) [ 647.409039] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.409039] env[63175]: result = function(*args, **kwargs) [ 647.409039] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.409039] env[63175]: return func(*args, **kwargs) [ 647.409039] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.409039] env[63175]: raise e [ 647.409039] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.409039] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 647.409039] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.409039] env[63175]: created_port_ids = self._update_ports_for_instance( [ 647.409039] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.409039] env[63175]: with excutils.save_and_reraise_exception(): [ 647.409039] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.409039] env[63175]: self.force_reraise() [ 647.409039] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.409039] env[63175]: raise self.value [ 647.409039] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.409039] env[63175]: updated_port = self._update_port( [ 647.409039] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.409039] env[63175]: _ensure_no_port_binding_failure(port) [ 647.409039] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.409039] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.409810] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 647.409810] env[63175]: Removing descriptor: 17 [ 647.424300] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.424540] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.424698] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.424886] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.425043] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.425199] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.425404] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.425560] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.425723] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.425882] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.426092] env[63175]: DEBUG nova.virt.hardware [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.426921] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b894a3c-844b-4156-9ab6-bb8612a8f841 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.437871] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abeb2ae8-a7c8-4fdb-9acd-693aecb4dc0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.451358] env[63175]: ERROR nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Traceback (most recent call last): [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] yield resources [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.driver.spawn(context, instance, image_meta, [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] vm_ref = self.build_virtual_machine(instance, [ 647.451358] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] for vif in network_info: [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return self._sync_wrapper(fn, *args, **kwargs) [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.wait() [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self[:] = self._gt.wait() [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return self._exit_event.wait() [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 647.451701] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] current.throw(*self._exc) [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] result = function(*args, **kwargs) [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return func(*args, **kwargs) [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise e [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] nwinfo = self.network_api.allocate_for_instance( [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] created_port_ids = self._update_ports_for_instance( [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] with excutils.save_and_reraise_exception(): [ 647.452070] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.force_reraise() [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise self.value [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] updated_port = self._update_port( [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] _ensure_no_port_binding_failure(port) [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise exception.PortBindingFailed(port_id=port['id']) [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 647.452453] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] [ 647.452453] env[63175]: INFO nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Terminating instance [ 647.854011] env[63175]: INFO nova.compute.manager [-] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Took 1.06 seconds to deallocate network for instance. [ 647.855279] env[63175]: DEBUG nova.compute.claims [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.855645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.865440] env[63175]: DEBUG oslo_concurrency.lockutils [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] Releasing lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.865952] env[63175]: DEBUG nova.compute.manager [req-359f8d44-3f30-4146-86f6-56198af797e9 req-b78fd0a0-d4b6-4d4b-9a98-8384239f65c4 service nova] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Received event network-vif-deleted-fb241359-1ae8-4e7f-93ad-099af3f20e36 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 647.921056] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.947260] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.947491] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.958618] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquiring lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.958618] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquired lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.958618] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.032397] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.197900] env[63175]: DEBUG nova.compute.manager [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Received event network-changed-6c88289f-bce6-4bda-a058-f6913ebf1a02 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 648.198061] env[63175]: DEBUG nova.compute.manager [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Refreshing instance network info cache due to event network-changed-6c88289f-bce6-4bda-a058-f6913ebf1a02. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 648.198205] env[63175]: DEBUG oslo_concurrency.lockutils [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] Acquiring lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.302665] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25f1cc6-07ae-4e4c-b954-8a5ebcc7dc96 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.313175] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98fcd0e-2d24-48f1-9cdf-d2c7039b6ead {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.346849] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc469d58-3632-42d3-ac02-2221caae3736 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.354383] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c1ddde-52c8-47b3-ab62-6b6237b601c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.368481] env[63175]: DEBUG nova.compute.provider_tree [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.483561] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.539140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-1939e74b-3ab5-428b-86c5-4d6de694c454" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.539494] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 648.540168] env[63175]: DEBUG nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 648.540168] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.554836] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.568402] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.871101] env[63175]: DEBUG nova.scheduler.client.report [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 649.058591] env[63175]: DEBUG nova.network.neutron [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.071053] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Releasing lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.071497] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 649.071807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.072323] env[63175]: DEBUG oslo_concurrency.lockutils [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] Acquired lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.072445] env[63175]: DEBUG nova.network.neutron [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Refreshing network info cache for port 6c88289f-bce6-4bda-a058-f6913ebf1a02 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.073933] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b8c6b68-eb6b-454d-9e73-1b5f9a4055a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.084920] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adc56ae-1c11-4f6c-8b7b-3bdbcbc0bbb5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.106015] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9e39ddc-e4a5-417d-a9be-905bfc3f5ead could not be found. [ 649.106246] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.106432] env[63175]: INFO nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Took 0.03 seconds to destroy the instance on the hypervisor. [ 649.106670] env[63175]: DEBUG oslo.service.loopingcall [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.106883] env[63175]: DEBUG nova.compute.manager [-] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 649.106976] env[63175]: DEBUG nova.network.neutron [-] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.121087] env[63175]: DEBUG nova.network.neutron [-] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.376765] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.377465] env[63175]: ERROR nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Traceback (most recent call last): [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.driver.spawn(context, instance, image_meta, [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] vm_ref = self.build_virtual_machine(instance, [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.377465] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] for vif in network_info: [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return self._sync_wrapper(fn, *args, **kwargs) [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.wait() [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self[:] = self._gt.wait() [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return self._exit_event.wait() [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] current.throw(*self._exc) [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.377787] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] result = function(*args, **kwargs) [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] return func(*args, **kwargs) [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise e [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] nwinfo = self.network_api.allocate_for_instance( [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] created_port_ids = self._update_ports_for_instance( [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] with excutils.save_and_reraise_exception(): [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] self.force_reraise() [ 649.378202] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise self.value [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] updated_port = self._update_port( [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] _ensure_no_port_binding_failure(port) [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] raise exception.PortBindingFailed(port_id=port['id']) [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] nova.exception.PortBindingFailed: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. [ 649.378564] env[63175]: ERROR nova.compute.manager [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] [ 649.378564] env[63175]: DEBUG nova.compute.utils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.379511] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.461s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.380982] env[63175]: INFO nova.compute.claims [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.383732] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Build of instance 6cd44d90-9092-483b-ab80-442e8f59435e was re-scheduled: Binding failed for port d052f968-ecf5-4a74-ad38-ac1267ea32f4, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 649.384173] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 649.384401] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquiring lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.384550] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Acquired lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.384707] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.561085] env[63175]: INFO nova.compute.manager [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 1939e74b-3ab5-428b-86c5-4d6de694c454] Took 1.02 seconds to deallocate network for instance. [ 649.597265] env[63175]: DEBUG nova.network.neutron [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.623042] env[63175]: DEBUG nova.network.neutron [-] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.678179] env[63175]: DEBUG nova.network.neutron [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.902920] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.948136] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.125330] env[63175]: INFO nova.compute.manager [-] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Took 1.02 seconds to deallocate network for instance. [ 650.128601] env[63175]: DEBUG nova.compute.claims [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.128601] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.180052] env[63175]: DEBUG oslo_concurrency.lockutils [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] Releasing lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.180343] env[63175]: DEBUG nova.compute.manager [req-4992bc77-5832-484e-a6cb-fac43e71269f req-9c86fecd-e9c9-45e1-aad9-3c164bd49c82 service nova] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Received event network-vif-deleted-6c88289f-bce6-4bda-a058-f6913ebf1a02 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 650.451096] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Releasing lock "refresh_cache-6cd44d90-9092-483b-ab80-442e8f59435e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.451416] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 650.451589] env[63175]: DEBUG nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 650.451757] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.467620] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.590470] env[63175]: INFO nova.scheduler.client.report [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Deleted allocations for instance 1939e74b-3ab5-428b-86c5-4d6de694c454 [ 650.842357] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6ca52d-6e88-4678-8be7-c1f5832f83a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.850585] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b200e47f-bb99-4723-8fb8-52a618da693d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.881896] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715987fe-e429-48d5-adf7-8cd5d49c1400 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.888985] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fe6ff0-4eee-4c04-8d84-58cbefb73211 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.901944] env[63175]: DEBUG nova.compute.provider_tree [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.972721] env[63175]: DEBUG nova.network.neutron [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.101737] env[63175]: DEBUG oslo_concurrency.lockutils [None req-011c13a3-d290-42d6-8eb1-694daf11df44 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "1939e74b-3ab5-428b-86c5-4d6de694c454" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.585s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.405162] env[63175]: DEBUG nova.scheduler.client.report [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 651.475704] env[63175]: INFO nova.compute.manager [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] [instance: 6cd44d90-9092-483b-ab80-442e8f59435e] Took 1.02 seconds to deallocate network for instance. [ 651.606009] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 651.909954] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.910452] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 651.913106] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.080s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.123050] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.417444] env[63175]: DEBUG nova.compute.utils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.418918] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 652.419098] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.464662] env[63175]: DEBUG nova.policy [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6be51e1e838f4dc1b51e6a479a2ce560', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87426559e5124f61b553b06731e10314', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.499199] env[63175]: INFO nova.scheduler.client.report [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Deleted allocations for instance 6cd44d90-9092-483b-ab80-442e8f59435e [ 652.752523] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Successfully created port: 1713fe3b-7710-46c8-ad1b-aadcbab35ac1 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.789813] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d3a18a-ae18-41bf-aff7-aab1d6c8c4dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.798260] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e0f9ce-1c0d-471c-8c19-a76b38ac8f06 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.830243] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ccb192-4d5a-47f6-bf78-467d7bec51e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.837963] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df204680-971a-4eb2-b6de-3e074e278bf7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.850196] env[63175]: DEBUG nova.compute.provider_tree [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.924867] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 653.008702] env[63175]: DEBUG oslo_concurrency.lockutils [None req-52c5e871-1184-4e0a-b604-f2db8577a6aa tempest-DeleteServersAdminTestJSON-1784618456 tempest-DeleteServersAdminTestJSON-1784618456-project-member] Lock "6cd44d90-9092-483b-ab80-442e8f59435e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.216s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.356143] env[63175]: DEBUG nova.scheduler.client.report [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 653.511622] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 653.559244] env[63175]: DEBUG nova.compute.manager [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Received event network-changed-1713fe3b-7710-46c8-ad1b-aadcbab35ac1 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 653.559451] env[63175]: DEBUG nova.compute.manager [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Refreshing instance network info cache due to event network-changed-1713fe3b-7710-46c8-ad1b-aadcbab35ac1. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 653.559661] env[63175]: DEBUG oslo_concurrency.lockutils [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] Acquiring lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.559816] env[63175]: DEBUG oslo_concurrency.lockutils [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] Acquired lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.559978] env[63175]: DEBUG nova.network.neutron [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Refreshing network info cache for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.759669] env[63175]: ERROR nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 653.759669] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.759669] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.759669] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.759669] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.759669] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.759669] env[63175]: ERROR nova.compute.manager raise self.value [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.759669] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.759669] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.759669] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.760736] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.760736] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.760736] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 653.760736] env[63175]: ERROR nova.compute.manager [ 653.760736] env[63175]: Traceback (most recent call last): [ 653.760736] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.760736] env[63175]: listener.cb(fileno) [ 653.760736] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.760736] env[63175]: result = function(*args, **kwargs) [ 653.760736] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.760736] env[63175]: return func(*args, **kwargs) [ 653.760736] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 653.760736] env[63175]: raise e [ 653.760736] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.760736] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 653.760736] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.760736] env[63175]: created_port_ids = self._update_ports_for_instance( [ 653.760736] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.760736] env[63175]: with excutils.save_and_reraise_exception(): [ 653.760736] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.760736] env[63175]: self.force_reraise() [ 653.760736] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.760736] env[63175]: raise self.value [ 653.760736] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.760736] env[63175]: updated_port = self._update_port( [ 653.760736] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.760736] env[63175]: _ensure_no_port_binding_failure(port) [ 653.760736] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.760736] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.761481] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 653.761481] env[63175]: Removing descriptor: 17 [ 653.858960] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.859807] env[63175]: ERROR nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Traceback (most recent call last): [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.driver.spawn(context, instance, image_meta, [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] vm_ref = self.build_virtual_machine(instance, [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.859807] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] for vif in network_info: [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return self._sync_wrapper(fn, *args, **kwargs) [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.wait() [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self[:] = self._gt.wait() [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return self._exit_event.wait() [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] current.throw(*self._exc) [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.860149] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] result = function(*args, **kwargs) [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] return func(*args, **kwargs) [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise e [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] nwinfo = self.network_api.allocate_for_instance( [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] created_port_ids = self._update_ports_for_instance( [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] with excutils.save_and_reraise_exception(): [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] self.force_reraise() [ 653.860475] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise self.value [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] updated_port = self._update_port( [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] _ensure_no_port_binding_failure(port) [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] raise exception.PortBindingFailed(port_id=port['id']) [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] nova.exception.PortBindingFailed: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. [ 653.860837] env[63175]: ERROR nova.compute.manager [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] [ 653.860837] env[63175]: DEBUG nova.compute.utils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.861849] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.810s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.864717] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Build of instance ce379bc8-add6-4008-b4db-c8bdb61a8e97 was re-scheduled: Binding failed for port 9c863723-11d5-4577-8c1c-931b94e14451, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 653.865147] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 653.865366] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.865511] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.865667] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.934839] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 653.977296] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.977552] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.977709] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.977896] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.978088] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.978283] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.978470] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.978646] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.978839] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.979196] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.979196] env[63175]: DEBUG nova.virt.hardware [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.980133] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857e1f0f-0809-43b1-a4e4-fd0742887eeb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.987937] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07a342d-1c33-43be-bd14-591ed6351cb3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.001829] env[63175]: ERROR nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Traceback (most recent call last): [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] yield resources [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.driver.spawn(context, instance, image_meta, [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] vm_ref = self.build_virtual_machine(instance, [ 654.001829] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] for vif in network_info: [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return self._sync_wrapper(fn, *args, **kwargs) [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.wait() [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self[:] = self._gt.wait() [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return self._exit_event.wait() [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.002201] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] current.throw(*self._exc) [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] result = function(*args, **kwargs) [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return func(*args, **kwargs) [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise e [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] nwinfo = self.network_api.allocate_for_instance( [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] created_port_ids = self._update_ports_for_instance( [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] with excutils.save_and_reraise_exception(): [ 654.002599] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.force_reraise() [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise self.value [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] updated_port = self._update_port( [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] _ensure_no_port_binding_failure(port) [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise exception.PortBindingFailed(port_id=port['id']) [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 654.003028] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] [ 654.003028] env[63175]: INFO nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Terminating instance [ 654.037043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.078556] env[63175]: DEBUG nova.network.neutron [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.179969] env[63175]: DEBUG nova.network.neutron [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.394900] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.483187] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.507018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquiring lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.683687] env[63175]: DEBUG oslo_concurrency.lockutils [req-116b3356-8025-4217-b9bc-56de34f1cf32 req-2ef7a132-f778-424d-bb23-10f7e17ad358 service nova] Releasing lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.684152] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquired lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.684368] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.746977] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317cbebd-31cd-4f03-b763-ffc7ee10160c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.756021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1c83a0-a659-4d80-9db5-7502097cee80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.790632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8710090f-925b-4913-9864-e6ead24ff853 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.798126] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f9876e-4e3a-4143-905f-302bdcb55305 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.810977] env[63175]: DEBUG nova.compute.provider_tree [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.988683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-ce379bc8-add6-4008-b4db-c8bdb61a8e97" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.988803] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 654.988928] env[63175]: DEBUG nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 654.989154] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.004128] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.219341] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.318913] env[63175]: DEBUG nova.scheduler.client.report [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 655.326928] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.506312] env[63175]: DEBUG nova.network.neutron [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.598173] env[63175]: DEBUG nova.compute.manager [req-41956e61-cfe2-49ce-b098-745d16497015 req-3171e5c1-7ad7-41d2-925a-6e4c1f98a0fc service nova] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Received event network-vif-deleted-1713fe3b-7710-46c8-ad1b-aadcbab35ac1 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 655.821664] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.822820] env[63175]: ERROR nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Traceback (most recent call last): [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.driver.spawn(context, instance, image_meta, [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] vm_ref = self.build_virtual_machine(instance, [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.822820] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] for vif in network_info: [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return self._sync_wrapper(fn, *args, **kwargs) [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.wait() [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self[:] = self._gt.wait() [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return self._exit_event.wait() [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] current.throw(*self._exc) [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.823209] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] result = function(*args, **kwargs) [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] return func(*args, **kwargs) [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise e [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] nwinfo = self.network_api.allocate_for_instance( [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] created_port_ids = self._update_ports_for_instance( [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] with excutils.save_and_reraise_exception(): [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] self.force_reraise() [ 655.823526] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise self.value [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] updated_port = self._update_port( [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] _ensure_no_port_binding_failure(port) [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] raise exception.PortBindingFailed(port_id=port['id']) [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] nova.exception.PortBindingFailed: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. [ 655.823819] env[63175]: ERROR nova.compute.manager [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] [ 655.823819] env[63175]: DEBUG nova.compute.utils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.824565] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.264s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.828199] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Build of instance 425a1271-cda7-45e4-89cb-cccd6231b623 was re-scheduled: Binding failed for port 51cb423f-8eb2-47da-8647-453018aef00c, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 655.829025] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 655.829025] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquiring lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.829025] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Acquired lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.829242] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.830386] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Releasing lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.831094] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 655.831094] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.831219] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b37b4e3-2ead-4bdd-9107-9e38ed14e0e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.840518] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf58c12f-50d4-4b95-a5a3-335257b68436 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.866256] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b8b82813-b9c1-4d5c-964f-19016393b985 could not be found. [ 655.866566] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.866740] env[63175]: INFO nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.866983] env[63175]: DEBUG oslo.service.loopingcall [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.867748] env[63175]: DEBUG nova.compute.manager [-] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.867844] env[63175]: DEBUG nova.network.neutron [-] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.895584] env[63175]: DEBUG nova.network.neutron [-] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.009655] env[63175]: INFO nova.compute.manager [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: ce379bc8-add6-4008-b4db-c8bdb61a8e97] Took 1.02 seconds to deallocate network for instance. [ 656.353829] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.398051] env[63175]: DEBUG nova.network.neutron [-] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.484864] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.715127] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07e8589-8337-479e-ae69-fc328e724759 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.722945] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a98f96-64c7-494f-a458-a1fdf644d29c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.754610] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f79cf59-7634-44a0-a964-e4249932540b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.762121] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d31545e-356d-447e-ad2d-3feb58a356de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.782140] env[63175]: DEBUG nova.compute.provider_tree [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.901588] env[63175]: INFO nova.compute.manager [-] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Took 1.03 seconds to deallocate network for instance. [ 656.903955] env[63175]: DEBUG nova.compute.claims [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.904168] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.987443] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Releasing lock "refresh_cache-425a1271-cda7-45e4-89cb-cccd6231b623" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.987680] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 656.987859] env[63175]: DEBUG nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 656.988052] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.002736] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.045329] env[63175]: INFO nova.scheduler.client.report [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Deleted allocations for instance ce379bc8-add6-4008-b4db-c8bdb61a8e97 [ 657.290026] env[63175]: DEBUG nova.scheduler.client.report [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 657.508624] env[63175]: DEBUG nova.network.neutron [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.556197] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5e26377-fc0a-4630-9d82-b43908ef7339 tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "ce379bc8-add6-4008-b4db-c8bdb61a8e97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.319s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.797122] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.797122] env[63175]: ERROR nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Traceback (most recent call last): [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.driver.spawn(context, instance, image_meta, [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.797122] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] vm_ref = self.build_virtual_machine(instance, [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] for vif in network_info: [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return self._sync_wrapper(fn, *args, **kwargs) [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.wait() [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self[:] = self._gt.wait() [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return self._exit_event.wait() [ 657.797429] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] current.throw(*self._exc) [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] result = function(*args, **kwargs) [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] return func(*args, **kwargs) [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise e [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] nwinfo = self.network_api.allocate_for_instance( [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] created_port_ids = self._update_ports_for_instance( [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.797790] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] with excutils.save_and_reraise_exception(): [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] self.force_reraise() [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise self.value [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] updated_port = self._update_port( [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] _ensure_no_port_binding_failure(port) [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] raise exception.PortBindingFailed(port_id=port['id']) [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] nova.exception.PortBindingFailed: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. [ 657.798314] env[63175]: ERROR nova.compute.manager [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] [ 657.798654] env[63175]: DEBUG nova.compute.utils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.798654] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.207s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.804072] env[63175]: INFO nova.compute.claims [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.804072] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Build of instance 5da92241-84f0-4510-b19d-b28cb57079ab was re-scheduled: Binding failed for port f42a568c-7569-41be-8170-db96e586cf1e, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 657.804072] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 657.804554] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquiring lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.804802] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Acquired lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.805076] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.011229] env[63175]: INFO nova.compute.manager [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] [instance: 425a1271-cda7-45e4-89cb-cccd6231b623] Took 1.02 seconds to deallocate network for instance. [ 658.061556] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 658.326956] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.419330] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.592967] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.921553] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Releasing lock "refresh_cache-5da92241-84f0-4510-b19d-b28cb57079ab" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.921782] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 658.922224] env[63175]: DEBUG nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 658.922224] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.938976] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.050671] env[63175]: INFO nova.scheduler.client.report [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Deleted allocations for instance 425a1271-cda7-45e4-89cb-cccd6231b623 [ 659.233262] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11b99b1-4efd-45dc-88b6-325775926259 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.242364] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ea2e3d-fc5c-407e-a61e-f177f9603967 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.274243] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7374d482-5964-45c9-b8ab-205fe85c3487 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.281653] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3da1589-ad14-4e6a-9f33-680cb5f4dd3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.294590] env[63175]: DEBUG nova.compute.provider_tree [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.442014] env[63175]: DEBUG nova.network.neutron [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.566019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d160935-bd4c-4652-af2e-65bdc938eddd tempest-ImagesOneServerTestJSON-145416201 tempest-ImagesOneServerTestJSON-145416201-project-member] Lock "425a1271-cda7-45e4-89cb-cccd6231b623" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.078s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.798750] env[63175]: DEBUG nova.scheduler.client.report [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 659.944718] env[63175]: INFO nova.compute.manager [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] [instance: 5da92241-84f0-4510-b19d-b28cb57079ab] Took 1.02 seconds to deallocate network for instance. [ 660.066412] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 660.112084] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.112388] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.304685] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.305232] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 660.307801] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.887s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.587253] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.811606] env[63175]: DEBUG nova.compute.utils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.816650] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 660.816650] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.892876] env[63175]: DEBUG nova.policy [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '346fd19aa4674ebaa588b086eceb09de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b0ce95c54a240929c99f9a7740de356', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.977513] env[63175]: INFO nova.scheduler.client.report [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Deleted allocations for instance 5da92241-84f0-4510-b19d-b28cb57079ab [ 661.309391] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Successfully created port: d51c945e-81bd-4981-a1f4-812ae7e3631f {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.322158] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 661.351447] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.351447] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 2cb7696b-f146-4db6-9e83-385fb8af1127 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.351447] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance d9e39ddc-e4a5-417d-a9be-905bfc3f5ead actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.351447] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance b8b82813-b9c1-4d5c-964f-19016393b985 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.351589] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.490635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-924dc962-d6f8-48ad-a660-514f6d0f6ddb tempest-ListServerFiltersTestJSON-81411072 tempest-ListServerFiltersTestJSON-81411072-project-member] Lock "5da92241-84f0-4510-b19d-b28cb57079ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.068s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.853150] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 070514d0-ccc0-4fed-90bd-c7914760db09 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.994161] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 662.309533] env[63175]: DEBUG nova.compute.manager [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Received event network-changed-d51c945e-81bd-4981-a1f4-812ae7e3631f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 662.309886] env[63175]: DEBUG nova.compute.manager [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Refreshing instance network info cache due to event network-changed-d51c945e-81bd-4981-a1f4-812ae7e3631f. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 662.310023] env[63175]: DEBUG oslo_concurrency.lockutils [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] Acquiring lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.310511] env[63175]: DEBUG oslo_concurrency.lockutils [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] Acquired lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.310511] env[63175]: DEBUG nova.network.neutron [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Refreshing network info cache for port d51c945e-81bd-4981-a1f4-812ae7e3631f {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.338845] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 662.362148] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a5d426c1-9f30-43a7-942e-06cbda2fce30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.377898] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.377898] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.377898] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.378109] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.378109] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.378315] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.378510] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.378681] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.378837] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.379488] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.379488] env[63175]: DEBUG nova.virt.hardware [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.380370] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f99413d-c6b4-4070-a9c3-9f1afe44d585 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.391428] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d3c4e5-d356-40f8-9103-710023a18c7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.524623] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.548419] env[63175]: ERROR nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 662.548419] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.548419] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.548419] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.548419] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.548419] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.548419] env[63175]: ERROR nova.compute.manager raise self.value [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.548419] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.548419] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.548419] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.549589] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.549589] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.549589] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 662.549589] env[63175]: ERROR nova.compute.manager [ 662.549589] env[63175]: Traceback (most recent call last): [ 662.549589] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.549589] env[63175]: listener.cb(fileno) [ 662.549589] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.549589] env[63175]: result = function(*args, **kwargs) [ 662.549589] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.549589] env[63175]: return func(*args, **kwargs) [ 662.549589] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.549589] env[63175]: raise e [ 662.549589] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.549589] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 662.549589] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.549589] env[63175]: created_port_ids = self._update_ports_for_instance( [ 662.549589] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.549589] env[63175]: with excutils.save_and_reraise_exception(): [ 662.549589] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.549589] env[63175]: self.force_reraise() [ 662.549589] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.549589] env[63175]: raise self.value [ 662.549589] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.549589] env[63175]: updated_port = self._update_port( [ 662.549589] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.549589] env[63175]: _ensure_no_port_binding_failure(port) [ 662.549589] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.549589] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.550957] env[63175]: nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 662.550957] env[63175]: Removing descriptor: 17 [ 662.550957] env[63175]: ERROR nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Traceback (most recent call last): [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] yield resources [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.driver.spawn(context, instance, image_meta, [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.550957] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] vm_ref = self.build_virtual_machine(instance, [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] for vif in network_info: [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self._sync_wrapper(fn, *args, **kwargs) [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.wait() [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self[:] = self._gt.wait() [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self._exit_event.wait() [ 662.552109] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] result = hub.switch() [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self.greenlet.switch() [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] result = function(*args, **kwargs) [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return func(*args, **kwargs) [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise e [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] nwinfo = self.network_api.allocate_for_instance( [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.555408] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] created_port_ids = self._update_ports_for_instance( [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] with excutils.save_and_reraise_exception(): [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.force_reraise() [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise self.value [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] updated_port = self._update_port( [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] _ensure_no_port_binding_failure(port) [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.555862] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise exception.PortBindingFailed(port_id=port['id']) [ 662.556158] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 662.556158] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] [ 662.556158] env[63175]: INFO nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Terminating instance [ 662.843882] env[63175]: DEBUG nova.network.neutron [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.867160] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 7a25b67c-47b8-420f-9aa2-8c296b8ab221 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.941926] env[63175]: DEBUG nova.network.neutron [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.056244] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquiring lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.370609] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance d85e98e9-d3d5-4781-8428-d8ab517be146 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.447012] env[63175]: DEBUG oslo_concurrency.lockutils [req-bee60eb8-5cc2-4ada-83bf-00c5d7b3dec9 req-38bd413b-2430-45f3-92fe-d9b3b7ae3eb9 service nova] Releasing lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.447012] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquired lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.447012] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.874540] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 11747695-c79c-477d-b5ae-44c49dfb4bba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.970784] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.120794] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.335217] env[63175]: DEBUG nova.compute.manager [req-a9ff2c7c-4568-4586-bdac-dadf64808f8c req-9ab5c684-4dd5-4790-aa15-9fccaed5a082 service nova] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Received event network-vif-deleted-d51c945e-81bd-4981-a1f4-812ae7e3631f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 664.378155] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 8b305bcb-87e0-4fc9-9579-56a5af9ace8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.565354] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.565354] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.625384] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Releasing lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.625384] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 664.625384] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.625384] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e68b513-55d1-4483-8607-c142f543f46d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.635379] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f06fe8-96c0-4b20-8647-6c932e45d010 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.665024] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3 could not be found. [ 664.665024] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.665024] env[63175]: INFO nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 664.665024] env[63175]: DEBUG oslo.service.loopingcall [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.665024] env[63175]: DEBUG nova.compute.manager [-] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 664.665024] env[63175]: DEBUG nova.network.neutron [-] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.689529] env[63175]: DEBUG nova.network.neutron [-] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.882732] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 182b1cc6-9626-4146-be9a-e15c5690aa1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.193684] env[63175]: DEBUG nova.network.neutron [-] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.385282] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 8aaf9c9f-e992-49ca-acad-d518503544e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.696487] env[63175]: INFO nova.compute.manager [-] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Took 1.03 seconds to deallocate network for instance. [ 665.699405] env[63175]: DEBUG nova.compute.claims [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.699405] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.889908] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 223ccdbc-cee3-4040-a1ee-a53929025926 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.115847] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "a2e70590-17ed-4804-b232-57526e87d22b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.115847] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.394421] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance d5bc39a7-a493-41a3-a2f4-8958883adca8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.897046] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9af5d212-c032-4446-b3cd-87fe4e66ad7c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.400663] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 295e958d-20d4-493c-856b-71880cce8b98 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.904235] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 4dcb8813-c428-4e0f-a761-3b06a8cba173 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.406803] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.911998] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 501f803b-f4ea-4777-909e-7bb808628cc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.413830] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.916874] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 12a2d788-c995-468b-b2a2-17dba8ad01db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.420321] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e137904f-dc43-4ebb-90ab-e10ea5487fe5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.924599] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0f8e580e-fb14-4db8-b995-a9ffe06d8bac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.427095] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 67e8715d-b729-4013-8cca-44eaa55c662e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.930481] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 36089589-d105-49e0-8ae7-790c814b036c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.435696] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 39e55060-73a7-4dbd-96cf-bc48d8737c1c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.938257] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance c445245b-b7d3-49c6-82c5-1e8188c89b68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.938518] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 672.938674] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 673.283216] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c85fb2-5b6a-4ce0-bf33-f0959dc77688 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.290762] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bd081e-cf78-4a2f-9b70-294d312746c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.320213] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5714332b-bb04-41e2-a525-85edcfddf516 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.327202] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51903734-3646-4f7e-9c84-b0b4662ff527 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.339607] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.843049] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 674.346791] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 674.347018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.039s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.347314] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.913s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.189991] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cbf316-b3fc-4706-8645-68a96b93028a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.198124] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a23720-b26a-44a1-a489-427ff44cba06 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.229361] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d785307f-9bc2-4e28-90ec-3346471f1d33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.236807] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37bd1cb-c1dc-4493-b665-b2480b04b252 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.250843] env[63175]: DEBUG nova.compute.provider_tree [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.753427] env[63175]: DEBUG nova.scheduler.client.report [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 676.258697] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.259259] env[63175]: ERROR nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Traceback (most recent call last): [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.driver.spawn(context, instance, image_meta, [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] vm_ref = self.build_virtual_machine(instance, [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.259259] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] for vif in network_info: [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return self._sync_wrapper(fn, *args, **kwargs) [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.wait() [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self[:] = self._gt.wait() [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return self._exit_event.wait() [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] current.throw(*self._exc) [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.259605] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] result = function(*args, **kwargs) [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] return func(*args, **kwargs) [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise e [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] nwinfo = self.network_api.allocate_for_instance( [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] created_port_ids = self._update_ports_for_instance( [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] with excutils.save_and_reraise_exception(): [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] self.force_reraise() [ 676.259972] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise self.value [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] updated_port = self._update_port( [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] _ensure_no_port_binding_failure(port) [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] raise exception.PortBindingFailed(port_id=port['id']) [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] nova.exception.PortBindingFailed: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. [ 676.260399] env[63175]: ERROR nova.compute.manager [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] [ 676.260399] env[63175]: DEBUG nova.compute.utils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.261286] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.406s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.267020] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Build of instance 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa was re-scheduled: Binding failed for port e988c082-794b-4c1d-b983-e69778caf0a4, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 676.267020] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 676.267020] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.267020] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.267327] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.790250] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.923229] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.196106] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffef1b96-4bbb-4322-8480-17c86fd31959 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.203790] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9ecf98-aa30-407f-8ae9-6977859e868b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.234474] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1898e73-baa2-4e5c-a72e-7a7c06aa35af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.242490] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8747bb1e-1cfc-418f-8866-467626fdff41 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.256800] env[63175]: DEBUG nova.compute.provider_tree [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.425165] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.425429] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 677.425663] env[63175]: DEBUG nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 677.425844] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.455902] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.760663] env[63175]: DEBUG nova.scheduler.client.report [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 677.958684] env[63175]: DEBUG nova.network.neutron [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.265610] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.265788] env[63175]: ERROR nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Traceback (most recent call last): [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.driver.spawn(context, instance, image_meta, [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] vm_ref = self.build_virtual_machine(instance, [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.265788] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] for vif in network_info: [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self._sync_wrapper(fn, *args, **kwargs) [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.wait() [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self[:] = self._gt.wait() [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self._exit_event.wait() [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] result = hub.switch() [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.266112] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return self.greenlet.switch() [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] result = function(*args, **kwargs) [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] return func(*args, **kwargs) [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise e [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] nwinfo = self.network_api.allocate_for_instance( [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] created_port_ids = self._update_ports_for_instance( [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] with excutils.save_and_reraise_exception(): [ 678.266434] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] self.force_reraise() [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise self.value [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] updated_port = self._update_port( [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] _ensure_no_port_binding_failure(port) [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] raise exception.PortBindingFailed(port_id=port['id']) [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] nova.exception.PortBindingFailed: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. [ 678.266830] env[63175]: ERROR nova.compute.manager [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] [ 678.267121] env[63175]: DEBUG nova.compute.utils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 678.267686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.139s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.270862] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Build of instance 2cb7696b-f146-4db6-9e83-385fb8af1127 was re-scheduled: Binding failed for port fb241359-1ae8-4e7f-93ad-099af3f20e36, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 678.271067] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 678.271293] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquiring lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.271440] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Acquired lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.271616] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.461710] env[63175]: INFO nova.compute.manager [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa] Took 1.04 seconds to deallocate network for instance. [ 678.793983] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.880429] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.112882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919cb6da-2664-4ecb-869b-18b4548b0719 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.120176] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d21f35a-1410-4f7e-bb7a-22b21ca30256 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.150876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6571bb42-28db-4146-a998-7390f8e5abeb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.157740] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de448dbe-b70f-49e0-bd36-f587bd2eb2f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.170955] env[63175]: DEBUG nova.compute.provider_tree [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.386330] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Releasing lock "refresh_cache-2cb7696b-f146-4db6-9e83-385fb8af1127" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.386578] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 679.386766] env[63175]: DEBUG nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 679.386937] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.407136] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.491885] env[63175]: INFO nova.scheduler.client.report [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocations for instance 27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa [ 679.674336] env[63175]: DEBUG nova.scheduler.client.report [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 679.910614] env[63175]: DEBUG nova.network.neutron [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.001151] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69bcd459-f856-4383-aaf6-b66dd3dc331d tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "27f2e664-e3a4-4d12-8a9a-a7b8f7ff2caa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.281s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.178821] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.180038] env[63175]: ERROR nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Traceback (most recent call last): [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.driver.spawn(context, instance, image_meta, [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] vm_ref = self.build_virtual_machine(instance, [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.180038] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] for vif in network_info: [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return self._sync_wrapper(fn, *args, **kwargs) [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.wait() [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self[:] = self._gt.wait() [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return self._exit_event.wait() [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] current.throw(*self._exc) [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.180418] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] result = function(*args, **kwargs) [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] return func(*args, **kwargs) [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise e [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] nwinfo = self.network_api.allocate_for_instance( [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] created_port_ids = self._update_ports_for_instance( [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] with excutils.save_and_reraise_exception(): [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] self.force_reraise() [ 680.180747] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise self.value [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] updated_port = self._update_port( [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] _ensure_no_port_binding_failure(port) [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] raise exception.PortBindingFailed(port_id=port['id']) [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] nova.exception.PortBindingFailed: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. [ 680.181063] env[63175]: ERROR nova.compute.manager [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] [ 680.181063] env[63175]: DEBUG nova.compute.utils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.182275] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.059s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.183522] env[63175]: INFO nova.compute.claims [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.186249] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Build of instance d9e39ddc-e4a5-417d-a9be-905bfc3f5ead was re-scheduled: Binding failed for port 6c88289f-bce6-4bda-a058-f6913ebf1a02, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 680.186686] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 680.186911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquiring lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.187067] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Acquired lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.187334] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.414396] env[63175]: INFO nova.compute.manager [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] [instance: 2cb7696b-f146-4db6-9e83-385fb8af1127] Took 1.03 seconds to deallocate network for instance. [ 680.504686] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 680.705965] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.815688] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.021705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.302217] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.302452] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.317767] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Releasing lock "refresh_cache-d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.317989] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 681.318184] env[63175]: DEBUG nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 681.318358] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.333636] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.437147] env[63175]: INFO nova.scheduler.client.report [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Deleted allocations for instance 2cb7696b-f146-4db6-9e83-385fb8af1127 [ 681.523612] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d44699-2626-48c5-bb16-aa69c2b674a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.531736] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614b747c-71f2-425a-b7d2-ac1e12b45b0c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.560965] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8abdb9-6812-434f-b646-6767e77f604f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.568870] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a665a8e-b69a-4e00-88a7-1e2b497fa630 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.582133] env[63175]: DEBUG nova.compute.provider_tree [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.838137] env[63175]: DEBUG nova.network.neutron [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.946474] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41367779-74a3-4b4d-8558-67cc910272c4 tempest-ServersTestFqdnHostnames-1652452562 tempest-ServersTestFqdnHostnames-1652452562-project-member] Lock "2cb7696b-f146-4db6-9e83-385fb8af1127" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.605s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.085966] env[63175]: DEBUG nova.scheduler.client.report [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 682.342204] env[63175]: INFO nova.compute.manager [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] [instance: d9e39ddc-e4a5-417d-a9be-905bfc3f5ead] Took 1.02 seconds to deallocate network for instance. [ 682.452518] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 682.591825] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.592144] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 682.594452] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.558s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.595784] env[63175]: INFO nova.compute.claims [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.975808] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.099974] env[63175]: DEBUG nova.compute.utils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.103468] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 683.103652] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 683.197481] env[63175]: DEBUG nova.policy [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c44e3eb4d2e473b952aadbbc9e50338', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7791eda987384331a57213ea754060cf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.374492] env[63175]: INFO nova.scheduler.client.report [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Deleted allocations for instance d9e39ddc-e4a5-417d-a9be-905bfc3f5ead [ 683.604100] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 683.800014] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Successfully created port: 04fc6cab-f27c-45da-bc49-68ec422110ea {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.888176] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9adff0f2-f033-4438-bdb2-50693fad4c73 tempest-InstanceActionsV221TestJSON-261867782 tempest-InstanceActionsV221TestJSON-261867782-project-member] Lock "d9e39ddc-e4a5-417d-a9be-905bfc3f5ead" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.011s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.041031] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd09aa4f-0058-4693-9e6f-b2a1ffc88ec2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.045652] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd51de83-44fa-4057-b7ad-4f961e1dc43f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.076177] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4cd5a2-2c0a-439a-90b8-f99989f52787 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.083991] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3051060-3e3f-42f2-b983-66882653c6ef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.097760] env[63175]: DEBUG nova.compute.provider_tree [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.391905] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 684.604099] env[63175]: DEBUG nova.scheduler.client.report [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 684.616233] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 684.650238] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.650501] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.653123] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.653123] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.653123] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.653123] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.653123] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.653417] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.653417] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.653417] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.653417] env[63175]: DEBUG nova.virt.hardware [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.653909] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f721f5b-34b9-4208-82a9-e343b70aa2ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.663882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab17d44-1c3d-4d46-9b11-d234bb1b8290 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.759439] env[63175]: DEBUG nova.compute.manager [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Received event network-changed-04fc6cab-f27c-45da-bc49-68ec422110ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 684.759439] env[63175]: DEBUG nova.compute.manager [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Refreshing instance network info cache due to event network-changed-04fc6cab-f27c-45da-bc49-68ec422110ea. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 684.759439] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] Acquiring lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.759439] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] Acquired lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.759439] env[63175]: DEBUG nova.network.neutron [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Refreshing network info cache for port 04fc6cab-f27c-45da-bc49-68ec422110ea {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.918112] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.040616] env[63175]: ERROR nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 685.040616] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.040616] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.040616] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.040616] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.040616] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.040616] env[63175]: ERROR nova.compute.manager raise self.value [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.040616] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.040616] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.040616] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.041144] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.041144] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.041144] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 685.041144] env[63175]: ERROR nova.compute.manager [ 685.041144] env[63175]: Traceback (most recent call last): [ 685.041144] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.041144] env[63175]: listener.cb(fileno) [ 685.041144] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.041144] env[63175]: result = function(*args, **kwargs) [ 685.041144] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.041144] env[63175]: return func(*args, **kwargs) [ 685.041144] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 685.041144] env[63175]: raise e [ 685.041144] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.041144] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 685.041144] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.041144] env[63175]: created_port_ids = self._update_ports_for_instance( [ 685.041144] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.041144] env[63175]: with excutils.save_and_reraise_exception(): [ 685.041144] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.041144] env[63175]: self.force_reraise() [ 685.041144] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.041144] env[63175]: raise self.value [ 685.041144] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.041144] env[63175]: updated_port = self._update_port( [ 685.041144] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.041144] env[63175]: _ensure_no_port_binding_failure(port) [ 685.041144] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.041144] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.042864] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 685.042864] env[63175]: Removing descriptor: 17 [ 685.042864] env[63175]: ERROR nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Traceback (most recent call last): [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] yield resources [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.driver.spawn(context, instance, image_meta, [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.042864] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] vm_ref = self.build_virtual_machine(instance, [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] for vif in network_info: [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self._sync_wrapper(fn, *args, **kwargs) [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.wait() [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self[:] = self._gt.wait() [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self._exit_event.wait() [ 685.043226] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] result = hub.switch() [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self.greenlet.switch() [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] result = function(*args, **kwargs) [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return func(*args, **kwargs) [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise e [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] nwinfo = self.network_api.allocate_for_instance( [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.043653] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] created_port_ids = self._update_ports_for_instance( [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] with excutils.save_and_reraise_exception(): [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.force_reraise() [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise self.value [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] updated_port = self._update_port( [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] _ensure_no_port_binding_failure(port) [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.044031] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise exception.PortBindingFailed(port_id=port['id']) [ 685.044470] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 685.044470] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] [ 685.044470] env[63175]: INFO nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Terminating instance [ 685.109561] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.110087] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 685.114292] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.210s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.277947] env[63175]: DEBUG nova.network.neutron [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.357574] env[63175]: DEBUG nova.network.neutron [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.549221] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquiring lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.618744] env[63175]: DEBUG nova.compute.utils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.625097] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 685.625097] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 685.666902] env[63175]: DEBUG nova.policy [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b3629f6f704427ab8742e6b3cd61231', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f25949ff9c9944dfb1b442c917da864b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.860975] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b9aae19-f7d2-4eb7-a929-7d409ce670cd req-acbb17a4-26e4-43eb-bcdd-dd6b77a1148e service nova] Releasing lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.861686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquired lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.862066] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.070030] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f0489d-fb6e-468b-8cf4-ad5263f268fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.077830] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Successfully created port: 761de51a-b058-4574-b5c7-512c29b1f8a2 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.081105] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2244e09f-2d52-4b03-a743-7da2b91d0f73 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.111753] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c093160-f86d-47ca-8070-917dea3e1915 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.119420] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be81ed2-bb7e-4da3-9c65-a7b74de805f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.124040] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 686.138054] env[63175]: DEBUG nova.compute.provider_tree [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.382801] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.505761] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.644101] env[63175]: DEBUG nova.scheduler.client.report [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 686.882146] env[63175]: DEBUG nova.compute.manager [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Received event network-vif-deleted-04fc6cab-f27c-45da-bc49-68ec422110ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 686.882378] env[63175]: DEBUG nova.compute.manager [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Received event network-changed-761de51a-b058-4574-b5c7-512c29b1f8a2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 686.882534] env[63175]: DEBUG nova.compute.manager [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Refreshing instance network info cache due to event network-changed-761de51a-b058-4574-b5c7-512c29b1f8a2. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 686.882754] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] Acquiring lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.882893] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] Acquired lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.883067] env[63175]: DEBUG nova.network.neutron [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Refreshing network info cache for port 761de51a-b058-4574-b5c7-512c29b1f8a2 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.008890] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Releasing lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.009336] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 687.009530] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.009828] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71ef4f26-351a-4fc0-97b8-1d49ad8c5079 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.018809] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320ca514-f4d8-42bf-aae6-9ec1d5a7ba3e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.041408] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 070514d0-ccc0-4fed-90bd-c7914760db09 could not be found. [ 687.041637] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.041817] env[63175]: INFO nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.042074] env[63175]: DEBUG oslo.service.loopingcall [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.042316] env[63175]: DEBUG nova.compute.manager [-] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 687.042421] env[63175]: DEBUG nova.network.neutron [-] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.055901] env[63175]: ERROR nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 687.055901] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.055901] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.055901] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.055901] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.055901] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.055901] env[63175]: ERROR nova.compute.manager raise self.value [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.055901] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.055901] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.055901] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.056603] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.056603] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.056603] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 687.056603] env[63175]: ERROR nova.compute.manager [ 687.056603] env[63175]: Traceback (most recent call last): [ 687.056603] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.056603] env[63175]: listener.cb(fileno) [ 687.056603] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.056603] env[63175]: result = function(*args, **kwargs) [ 687.056603] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.056603] env[63175]: return func(*args, **kwargs) [ 687.056603] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.056603] env[63175]: raise e [ 687.056603] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.056603] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 687.056603] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.056603] env[63175]: created_port_ids = self._update_ports_for_instance( [ 687.056603] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.056603] env[63175]: with excutils.save_and_reraise_exception(): [ 687.056603] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.056603] env[63175]: self.force_reraise() [ 687.056603] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.056603] env[63175]: raise self.value [ 687.056603] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.056603] env[63175]: updated_port = self._update_port( [ 687.056603] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.056603] env[63175]: _ensure_no_port_binding_failure(port) [ 687.056603] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.056603] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.057779] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 687.057779] env[63175]: Removing descriptor: 17 [ 687.060749] env[63175]: DEBUG nova.network.neutron [-] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.149082] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 687.152078] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.152650] env[63175]: ERROR nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Traceback (most recent call last): [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.driver.spawn(context, instance, image_meta, [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] vm_ref = self.build_virtual_machine(instance, [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.152650] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] for vif in network_info: [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return self._sync_wrapper(fn, *args, **kwargs) [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.wait() [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self[:] = self._gt.wait() [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return self._exit_event.wait() [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] current.throw(*self._exc) [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.152954] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] result = function(*args, **kwargs) [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] return func(*args, **kwargs) [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise e [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] nwinfo = self.network_api.allocate_for_instance( [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] created_port_ids = self._update_ports_for_instance( [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] with excutils.save_and_reraise_exception(): [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] self.force_reraise() [ 687.153374] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise self.value [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] updated_port = self._update_port( [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] _ensure_no_port_binding_failure(port) [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] raise exception.PortBindingFailed(port_id=port['id']) [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] nova.exception.PortBindingFailed: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. [ 687.153758] env[63175]: ERROR nova.compute.manager [instance: b8b82813-b9c1-4d5c-964f-19016393b985] [ 687.153758] env[63175]: DEBUG nova.compute.utils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.154761] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.562s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.156324] env[63175]: INFO nova.compute.claims [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.159031] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Build of instance b8b82813-b9c1-4d5c-964f-19016393b985 was re-scheduled: Binding failed for port 1713fe3b-7710-46c8-ad1b-aadcbab35ac1, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 687.163281] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 687.163281] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquiring lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.163281] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Acquired lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.163281] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.178740] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.179119] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.179300] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.179489] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.179636] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.179778] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.179980] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.180151] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.180316] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.180478] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.180681] env[63175]: DEBUG nova.virt.hardware [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.181816] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dac926-4d0a-4de4-b4e9-379289a6cc18 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.195029] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70053d6-d1e5-4484-929e-778e72027251 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.209505] env[63175]: ERROR nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Traceback (most recent call last): [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] yield resources [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.driver.spawn(context, instance, image_meta, [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] vm_ref = self.build_virtual_machine(instance, [ 687.209505] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] for vif in network_info: [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return self._sync_wrapper(fn, *args, **kwargs) [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.wait() [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self[:] = self._gt.wait() [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return self._exit_event.wait() [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.209858] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] current.throw(*self._exc) [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] result = function(*args, **kwargs) [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return func(*args, **kwargs) [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise e [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] nwinfo = self.network_api.allocate_for_instance( [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] created_port_ids = self._update_ports_for_instance( [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] with excutils.save_and_reraise_exception(): [ 687.210183] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.force_reraise() [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise self.value [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] updated_port = self._update_port( [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] _ensure_no_port_binding_failure(port) [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise exception.PortBindingFailed(port_id=port['id']) [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 687.210586] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] [ 687.210586] env[63175]: INFO nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Terminating instance [ 687.406639] env[63175]: DEBUG nova.network.neutron [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.554157] env[63175]: DEBUG nova.network.neutron [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.562953] env[63175]: DEBUG nova.network.neutron [-] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.713943] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquiring lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.745291] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.876355] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.058785] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b3dcd47-c0f6-4d5e-bf1e-5557b3e60fa1 req-b582d1ba-ea84-4c8b-b868-b79ef2b812af service nova] Releasing lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.060803] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquired lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.060900] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.066906] env[63175]: INFO nova.compute.manager [-] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Took 1.02 seconds to deallocate network for instance. [ 688.069332] env[63175]: DEBUG nova.compute.claims [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.069511] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.382689] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Releasing lock "refresh_cache-b8b82813-b9c1-4d5c-964f-19016393b985" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.386773] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 688.386773] env[63175]: DEBUG nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 688.386773] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.423799] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.558842] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83c30ab-aa4c-48ef-bf45-5c90a5198c5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.568125] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f8ae5e-cc21-4798-aa29-3fd3e4e08f03 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.609237] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.611613] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539ceccf-de9e-4608-8853-a33f02d24394 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.619602] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d05cb45-c723-443c-90e5-4ba5272c32af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.633296] env[63175]: DEBUG nova.compute.provider_tree [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.742704] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.904818] env[63175]: DEBUG nova.compute.manager [req-7a27e58a-cbf1-4fab-b505-b19d7364c01c req-5ceaebbc-7ff4-4c71-89a1-bada6336277a service nova] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Received event network-vif-deleted-761de51a-b058-4574-b5c7-512c29b1f8a2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 688.926806] env[63175]: DEBUG nova.network.neutron [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.137071] env[63175]: DEBUG nova.scheduler.client.report [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 689.245297] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Releasing lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.245723] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 689.245905] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.246218] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36363869-8c37-4f99-90d2-6b1884d580dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.254864] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e801c01-252d-490c-a378-21bc7c8fd7cf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.276207] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a5d426c1-9f30-43a7-942e-06cbda2fce30 could not be found. [ 689.276425] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.276607] env[63175]: INFO nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Took 0.03 seconds to destroy the instance on the hypervisor. [ 689.276843] env[63175]: DEBUG oslo.service.loopingcall [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.277074] env[63175]: DEBUG nova.compute.manager [-] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 689.277170] env[63175]: DEBUG nova.network.neutron [-] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.292262] env[63175]: DEBUG nova.network.neutron [-] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.429689] env[63175]: INFO nova.compute.manager [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] [instance: b8b82813-b9c1-4d5c-964f-19016393b985] Took 1.04 seconds to deallocate network for instance. [ 689.642264] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.642786] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 689.645413] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.058s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.646818] env[63175]: INFO nova.compute.claims [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.795253] env[63175]: DEBUG nova.network.neutron [-] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.152226] env[63175]: DEBUG nova.compute.utils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.156023] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 690.156023] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.208427] env[63175]: DEBUG nova.policy [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db76f4c61f9a4aea80c546253b937b3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7126014d48214eda98fb43c1b774133f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.302063] env[63175]: INFO nova.compute.manager [-] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Took 1.02 seconds to deallocate network for instance. [ 690.304275] env[63175]: DEBUG nova.compute.claims [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.304510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.471893] env[63175]: INFO nova.scheduler.client.report [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Deleted allocations for instance b8b82813-b9c1-4d5c-964f-19016393b985 [ 690.622434] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Successfully created port: 08649e21-1bb9-4b4c-9b42-5254e44d28c2 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.656710] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 690.984849] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f256572d-3e45-45d7-87ed-68c0aa97b8b9 tempest-ImagesOneServerNegativeTestJSON-253820786 tempest-ImagesOneServerNegativeTestJSON-253820786-project-member] Lock "b8b82813-b9c1-4d5c-964f-19016393b985" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.938s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.044141] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d730fea3-1df8-41b0-9963-d070e165dc30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.052949] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8717ce0-ce48-4ee1-b5ce-d76a07206d2e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.085210] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544c3c97-7ad8-434f-8a31-fb48fccb3edb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.090093] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811a3bd1-f0d4-4a4a-ac0b-2699fa1b2ee3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.103631] env[63175]: DEBUG nova.compute.provider_tree [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.146424] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.146657] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.431853] env[63175]: DEBUG nova.compute.manager [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Received event network-changed-08649e21-1bb9-4b4c-9b42-5254e44d28c2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 691.432086] env[63175]: DEBUG nova.compute.manager [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Refreshing instance network info cache due to event network-changed-08649e21-1bb9-4b4c-9b42-5254e44d28c2. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 691.432394] env[63175]: DEBUG oslo_concurrency.lockutils [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] Acquiring lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.432394] env[63175]: DEBUG oslo_concurrency.lockutils [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] Acquired lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.432925] env[63175]: DEBUG nova.network.neutron [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Refreshing network info cache for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.490056] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 691.608186] env[63175]: DEBUG nova.scheduler.client.report [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 691.612907] env[63175]: ERROR nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 691.612907] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.612907] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.612907] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.612907] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.612907] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.612907] env[63175]: ERROR nova.compute.manager raise self.value [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.612907] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.612907] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.612907] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.613489] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.613489] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.613489] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 691.613489] env[63175]: ERROR nova.compute.manager [ 691.613489] env[63175]: Traceback (most recent call last): [ 691.613489] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.613489] env[63175]: listener.cb(fileno) [ 691.613489] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.613489] env[63175]: result = function(*args, **kwargs) [ 691.613489] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.613489] env[63175]: return func(*args, **kwargs) [ 691.613489] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.613489] env[63175]: raise e [ 691.613489] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.613489] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 691.613489] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.613489] env[63175]: created_port_ids = self._update_ports_for_instance( [ 691.613489] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.613489] env[63175]: with excutils.save_and_reraise_exception(): [ 691.613489] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.613489] env[63175]: self.force_reraise() [ 691.613489] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.613489] env[63175]: raise self.value [ 691.613489] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.613489] env[63175]: updated_port = self._update_port( [ 691.613489] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.613489] env[63175]: _ensure_no_port_binding_failure(port) [ 691.613489] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.613489] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.614183] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 691.614183] env[63175]: Removing descriptor: 15 [ 691.670724] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 691.706422] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.706675] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.707412] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.707412] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.707412] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.707412] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.707695] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.707862] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.708075] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.708276] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.708499] env[63175]: DEBUG nova.virt.hardware [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.709537] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed05731-ca18-4034-a43a-bb8a965737af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.718103] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9317a879-57e8-4754-bf7b-059defb20503 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.732342] env[63175]: ERROR nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Traceback (most recent call last): [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] yield resources [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.driver.spawn(context, instance, image_meta, [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] vm_ref = self.build_virtual_machine(instance, [ 691.732342] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] for vif in network_info: [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return self._sync_wrapper(fn, *args, **kwargs) [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.wait() [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self[:] = self._gt.wait() [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return self._exit_event.wait() [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.732817] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] current.throw(*self._exc) [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] result = function(*args, **kwargs) [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return func(*args, **kwargs) [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise e [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] nwinfo = self.network_api.allocate_for_instance( [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] created_port_ids = self._update_ports_for_instance( [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] with excutils.save_and_reraise_exception(): [ 691.733284] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.force_reraise() [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise self.value [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] updated_port = self._update_port( [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] _ensure_no_port_binding_failure(port) [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise exception.PortBindingFailed(port_id=port['id']) [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 691.733717] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] [ 691.733717] env[63175]: INFO nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Terminating instance [ 691.950693] env[63175]: DEBUG nova.network.neutron [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.014467] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.048601] env[63175]: DEBUG nova.network.neutron [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.115902] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.115902] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 692.118331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.596s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.120019] env[63175]: INFO nova.compute.claims [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.237864] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.549926] env[63175]: DEBUG oslo_concurrency.lockutils [req-f9486b47-26b0-412d-8665-b4e7217c89f0 req-10b1c315-92a7-4277-a808-e55112256059 service nova] Releasing lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.550362] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.550583] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.624182] env[63175]: DEBUG nova.compute.utils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.628551] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 692.628551] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 692.685449] env[63175]: DEBUG nova.policy [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '863baae1d2194eeebb4b59f72eded329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b34e1c3fa6347fcba22f391626741a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.084084] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.128227] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 693.182120] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.315608] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Successfully created port: 43932b33-1509-4d42-8684-f7093e2aebc9 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.466594] env[63175]: DEBUG nova.compute.manager [req-872c7f2e-4f31-4eb0-847f-aa76aeb69743 req-16e38877-0654-4798-a692-cd03cca85e28 service nova] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Received event network-vif-deleted-08649e21-1bb9-4b4c-9b42-5254e44d28c2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 693.555984] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8511e1c6-6288-4f9c-bd65-670355076dd8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.564056] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b42f86-e8ab-4fe4-ae8d-35bf032e1f41 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.593222] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9525cf-513b-4bbc-bc50-e4e634c7fb40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.600639] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ee0964-5b5b-4de3-a9d7-80dd598e2ff9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.613517] env[63175]: DEBUG nova.compute.provider_tree [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.687510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Releasing lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.688109] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 693.688212] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.689121] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-319a426f-40cb-46f5-96df-48e477d48ffb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.697632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f4bb32-5feb-451e-b6e7-6084e82049f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.719383] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a25b67c-47b8-420f-9aa2-8c296b8ab221 could not be found. [ 693.719630] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.719821] env[63175]: INFO nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Took 0.03 seconds to destroy the instance on the hypervisor. [ 693.720083] env[63175]: DEBUG oslo.service.loopingcall [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.720316] env[63175]: DEBUG nova.compute.manager [-] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 693.720412] env[63175]: DEBUG nova.network.neutron [-] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.735820] env[63175]: DEBUG nova.network.neutron [-] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.119069] env[63175]: DEBUG nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 694.142577] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 694.170044] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.170300] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.170456] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.170718] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.170876] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.171036] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.171253] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.171412] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.171596] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.171788] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.171962] env[63175]: DEBUG nova.virt.hardware [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.173074] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d1bf36-4ff5-4e0d-93e1-d0dad6d40d13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.181145] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4518cd78-f60f-4a8c-a8a9-12a8d4958dc7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.241918] env[63175]: DEBUG nova.network.neutron [-] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.551839] env[63175]: ERROR nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 694.551839] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.551839] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.551839] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.551839] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.551839] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.551839] env[63175]: ERROR nova.compute.manager raise self.value [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.551839] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.551839] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.551839] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.552531] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.552531] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.552531] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 694.552531] env[63175]: ERROR nova.compute.manager [ 694.552531] env[63175]: Traceback (most recent call last): [ 694.552531] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.552531] env[63175]: listener.cb(fileno) [ 694.552531] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.552531] env[63175]: result = function(*args, **kwargs) [ 694.552531] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.552531] env[63175]: return func(*args, **kwargs) [ 694.552531] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.552531] env[63175]: raise e [ 694.552531] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.552531] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 694.552531] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.552531] env[63175]: created_port_ids = self._update_ports_for_instance( [ 694.552531] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.552531] env[63175]: with excutils.save_and_reraise_exception(): [ 694.552531] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.552531] env[63175]: self.force_reraise() [ 694.552531] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.552531] env[63175]: raise self.value [ 694.552531] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.552531] env[63175]: updated_port = self._update_port( [ 694.552531] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.552531] env[63175]: _ensure_no_port_binding_failure(port) [ 694.552531] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.552531] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.555306] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 694.555306] env[63175]: Removing descriptor: 15 [ 694.555306] env[63175]: ERROR nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Traceback (most recent call last): [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] yield resources [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.driver.spawn(context, instance, image_meta, [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.555306] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] vm_ref = self.build_virtual_machine(instance, [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] for vif in network_info: [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self._sync_wrapper(fn, *args, **kwargs) [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.wait() [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self[:] = self._gt.wait() [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self._exit_event.wait() [ 694.555695] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] result = hub.switch() [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self.greenlet.switch() [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] result = function(*args, **kwargs) [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return func(*args, **kwargs) [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise e [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] nwinfo = self.network_api.allocate_for_instance( [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.556225] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] created_port_ids = self._update_ports_for_instance( [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] with excutils.save_and_reraise_exception(): [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.force_reraise() [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise self.value [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] updated_port = self._update_port( [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] _ensure_no_port_binding_failure(port) [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.556603] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise exception.PortBindingFailed(port_id=port['id']) [ 694.556986] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 694.556986] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] [ 694.556986] env[63175]: INFO nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Terminating instance [ 694.624782] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.624918] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 694.627360] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.928s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.745840] env[63175]: INFO nova.compute.manager [-] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Took 1.03 seconds to deallocate network for instance. [ 694.748208] env[63175]: DEBUG nova.compute.claims [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.748389] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.058818] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquiring lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.059043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquired lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.059231] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.131696] env[63175]: DEBUG nova.compute.utils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.136604] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 695.136786] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 695.203607] env[63175]: DEBUG nova.policy [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e3ed7f44704480380c91e92d7a94c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe51efa241a40a5ae4c306b1688ccae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.483035] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Successfully created port: de487c35-048b-4d84-9446-952e74b58d7b {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.500958] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de965515-4a3e-425c-bd59-359ee315b785 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.510215] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2b1eb2-2924-40de-b64e-fd75251ec952 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.542177] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653392db-4586-4c40-bd81-49396f5bb7a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.549322] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62346f53-ea95-4b21-9f22-363cd37b130b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.566237] env[63175]: DEBUG nova.compute.provider_tree [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.571350] env[63175]: DEBUG nova.compute.manager [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Received event network-changed-43932b33-1509-4d42-8684-f7093e2aebc9 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 695.571734] env[63175]: DEBUG nova.compute.manager [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Refreshing instance network info cache due to event network-changed-43932b33-1509-4d42-8684-f7093e2aebc9. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 695.571734] env[63175]: DEBUG oslo_concurrency.lockutils [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] Acquiring lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.580575] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.639873] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 695.677602] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.068871] env[63175]: DEBUG nova.scheduler.client.report [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 696.181347] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Releasing lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.181776] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 696.181970] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.182290] env[63175]: DEBUG oslo_concurrency.lockutils [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] Acquired lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.182463] env[63175]: DEBUG nova.network.neutron [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Refreshing network info cache for port 43932b33-1509-4d42-8684-f7093e2aebc9 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.183893] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28966666-ada6-4864-8282-2ef326f03bb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.192788] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98b7b7b-ac04-40d8-84e0-3a32f9256651 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.213739] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d85e98e9-d3d5-4781-8428-d8ab517be146 could not be found. [ 696.213957] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.214147] env[63175]: INFO nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Took 0.03 seconds to destroy the instance on the hypervisor. [ 696.214385] env[63175]: DEBUG oslo.service.loopingcall [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.214602] env[63175]: DEBUG nova.compute.manager [-] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.214696] env[63175]: DEBUG nova.network.neutron [-] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.230852] env[63175]: DEBUG nova.network.neutron [-] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.420966] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 696.420966] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 696.420966] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.420966] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.420966] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.420966] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.420966] env[63175]: ERROR nova.compute.manager raise self.value [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.420966] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 696.420966] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.420966] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 696.421628] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.421628] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 696.421628] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 696.421628] env[63175]: ERROR nova.compute.manager [ 696.421628] env[63175]: Traceback (most recent call last): [ 696.421628] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 696.421628] env[63175]: listener.cb(fileno) [ 696.421628] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.421628] env[63175]: result = function(*args, **kwargs) [ 696.421628] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.421628] env[63175]: return func(*args, **kwargs) [ 696.421628] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 696.421628] env[63175]: raise e [ 696.421628] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 696.421628] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 696.421628] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.421628] env[63175]: created_port_ids = self._update_ports_for_instance( [ 696.421628] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.421628] env[63175]: with excutils.save_and_reraise_exception(): [ 696.421628] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.421628] env[63175]: self.force_reraise() [ 696.421628] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.421628] env[63175]: raise self.value [ 696.421628] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.421628] env[63175]: updated_port = self._update_port( [ 696.421628] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.421628] env[63175]: _ensure_no_port_binding_failure(port) [ 696.421628] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.421628] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 696.422556] env[63175]: nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 696.422556] env[63175]: Removing descriptor: 15 [ 696.574013] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.574576] env[63175]: ERROR nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Traceback (most recent call last): [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.driver.spawn(context, instance, image_meta, [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] vm_ref = self.build_virtual_machine(instance, [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.574576] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] for vif in network_info: [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self._sync_wrapper(fn, *args, **kwargs) [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.wait() [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self[:] = self._gt.wait() [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self._exit_event.wait() [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] result = hub.switch() [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.575414] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return self.greenlet.switch() [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] result = function(*args, **kwargs) [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] return func(*args, **kwargs) [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise e [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] nwinfo = self.network_api.allocate_for_instance( [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] created_port_ids = self._update_ports_for_instance( [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] with excutils.save_and_reraise_exception(): [ 696.575752] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] self.force_reraise() [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise self.value [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] updated_port = self._update_port( [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] _ensure_no_port_binding_failure(port) [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] raise exception.PortBindingFailed(port_id=port['id']) [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] nova.exception.PortBindingFailed: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. [ 696.576134] env[63175]: ERROR nova.compute.manager [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] [ 696.576404] env[63175]: DEBUG nova.compute.utils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.576648] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.555s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.578036] env[63175]: INFO nova.compute.claims [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.582296] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Build of instance b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3 was re-scheduled: Binding failed for port d51c945e-81bd-4981-a1f4-812ae7e3631f, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 696.582296] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 696.582296] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquiring lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.582296] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Acquired lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.582514] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.648178] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 696.674077] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.674333] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.674485] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.674669] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.674816] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.675145] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.675411] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.676413] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.676413] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.676413] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.676413] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.676921] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa347ef-a714-47c7-bb47-53b925bb30bd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.685461] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadea062-bc3b-4652-9b32-77c4f1dc3eef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.703678] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Traceback (most recent call last): [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] yield resources [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.driver.spawn(context, instance, image_meta, [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] vm_ref = self.build_virtual_machine(instance, [ 696.703678] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] for vif in network_info: [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return self._sync_wrapper(fn, *args, **kwargs) [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.wait() [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self[:] = self._gt.wait() [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return self._exit_event.wait() [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.704080] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] current.throw(*self._exc) [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] result = function(*args, **kwargs) [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return func(*args, **kwargs) [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise e [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] nwinfo = self.network_api.allocate_for_instance( [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] created_port_ids = self._update_ports_for_instance( [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] with excutils.save_and_reraise_exception(): [ 696.704406] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.force_reraise() [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise self.value [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] updated_port = self._update_port( [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] _ensure_no_port_binding_failure(port) [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise exception.PortBindingFailed(port_id=port['id']) [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 696.704738] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] [ 696.704738] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Terminating instance [ 696.705984] env[63175]: DEBUG nova.network.neutron [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.733703] env[63175]: DEBUG nova.network.neutron [-] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.781143] env[63175]: DEBUG nova.network.neutron [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.101089] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.185825] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.210374] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.210543] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.210775] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.237440] env[63175]: INFO nova.compute.manager [-] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Took 1.02 seconds to deallocate network for instance. [ 697.237919] env[63175]: DEBUG nova.compute.claims [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.238105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.283336] env[63175]: DEBUG oslo_concurrency.lockutils [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] Releasing lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.283583] env[63175]: DEBUG nova.compute.manager [req-dcd7fa54-fa04-45f7-a4b4-1cd9ed737b5e req-611f74da-320c-4c04-9d2e-d135cd0e39e7 service nova] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Received event network-vif-deleted-43932b33-1509-4d42-8684-f7093e2aebc9 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 697.596179] env[63175]: DEBUG nova.compute.manager [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Received event network-changed-de487c35-048b-4d84-9446-952e74b58d7b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 697.596179] env[63175]: DEBUG nova.compute.manager [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Refreshing instance network info cache due to event network-changed-de487c35-048b-4d84-9446-952e74b58d7b. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 697.596179] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] Acquiring lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.688864] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Releasing lock "refresh_cache-b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.689122] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 697.689308] env[63175]: DEBUG nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 697.689472] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.704409] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.737695] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.812664] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.915503] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc5a993-fd2e-4f8c-a793-2c7073bceb55 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.923387] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07abfc6-0b99-41a1-a8c0-f1d906136b93 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.953990] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be47883c-a481-4fb7-8d17-88a6be5cba17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.961120] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d0dfd6-7b4b-451e-aa5e-93b8242aa20e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.976023] env[63175]: DEBUG nova.compute.provider_tree [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.208078] env[63175]: DEBUG nova.network.neutron [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.314861] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.315318] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 698.315508] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 698.315799] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] Acquired lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.315972] env[63175]: DEBUG nova.network.neutron [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Refreshing network info cache for port de487c35-048b-4d84-9446-952e74b58d7b {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.316973] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e956a3f9-8392-40dd-92f3-0a53ddd57d68 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.326277] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4e4951-69bc-450a-ba83-15aa02a6cb5c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.347562] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11747695-c79c-477d-b5ae-44c49dfb4bba could not be found. [ 698.347784] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.347967] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Took 0.03 seconds to destroy the instance on the hypervisor. [ 698.348222] env[63175]: DEBUG oslo.service.loopingcall [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.348447] env[63175]: DEBUG nova.compute.manager [-] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 698.348538] env[63175]: DEBUG nova.network.neutron [-] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.367204] env[63175]: DEBUG nova.network.neutron [-] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.478926] env[63175]: DEBUG nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 698.711025] env[63175]: INFO nova.compute.manager [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] [instance: b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3] Took 1.02 seconds to deallocate network for instance. [ 698.834326] env[63175]: DEBUG nova.network.neutron [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.869718] env[63175]: DEBUG nova.network.neutron [-] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.899798] env[63175]: DEBUG nova.network.neutron [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.983850] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.984402] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 698.986826] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.011s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.988612] env[63175]: INFO nova.compute.claims [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.372417] env[63175]: INFO nova.compute.manager [-] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Took 1.02 seconds to deallocate network for instance. [ 699.374863] env[63175]: DEBUG nova.compute.claims [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.375062] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.402130] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] Releasing lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.402382] env[63175]: DEBUG nova.compute.manager [req-8f416101-9f9a-40bb-9ec7-638dcd2420e0 req-5211066f-4dd7-4fc0-b168-92626528320c service nova] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Received event network-vif-deleted-de487c35-048b-4d84-9446-952e74b58d7b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 699.495527] env[63175]: DEBUG nova.compute.utils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.499365] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 699.499365] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.535553] env[63175]: DEBUG nova.policy [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e3ed7f44704480380c91e92d7a94c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe51efa241a40a5ae4c306b1688ccae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.744832] env[63175]: INFO nova.scheduler.client.report [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Deleted allocations for instance b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3 [ 699.838062] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Successfully created port: 5e7ae472-9404-4627-8af8-2536c5af0daf {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.999308] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 700.254833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-852d39e1-67f9-4cae-9442-0ae34ceeaaab tempest-ServerRescueTestJSONUnderV235-1094297641 tempest-ServerRescueTestJSONUnderV235-1094297641-project-member] Lock "b4acfb83-7200-46a3-9d0e-ae5d4ecb4dd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.448s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.332234] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0309f053-b232-491e-b9db-b9b8a509e139 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.340549] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d227cd8-c3d3-4a1c-9a1b-1fa2d9c232ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.369749] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862b8444-8eba-460d-a757-f5c180104b0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.377161] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385d0404-da17-41d6-abc1-41e9af9211b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.390261] env[63175]: DEBUG nova.compute.provider_tree [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.699919] env[63175]: DEBUG nova.compute.manager [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Received event network-changed-5e7ae472-9404-4627-8af8-2536c5af0daf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 700.700167] env[63175]: DEBUG nova.compute.manager [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Refreshing instance network info cache due to event network-changed-5e7ae472-9404-4627-8af8-2536c5af0daf. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 700.700389] env[63175]: DEBUG oslo_concurrency.lockutils [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] Acquiring lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.700530] env[63175]: DEBUG oslo_concurrency.lockutils [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] Acquired lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.700713] env[63175]: DEBUG nova.network.neutron [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Refreshing network info cache for port 5e7ae472-9404-4627-8af8-2536c5af0daf {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.759517] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 700.860200] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 700.860200] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 700.860200] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.860200] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.860200] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.860200] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.860200] env[63175]: ERROR nova.compute.manager raise self.value [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.860200] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 700.860200] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.860200] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 700.860629] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.860629] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 700.860629] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 700.860629] env[63175]: ERROR nova.compute.manager [ 700.860629] env[63175]: Traceback (most recent call last): [ 700.860629] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 700.860629] env[63175]: listener.cb(fileno) [ 700.860629] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.860629] env[63175]: result = function(*args, **kwargs) [ 700.860629] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 700.860629] env[63175]: return func(*args, **kwargs) [ 700.860629] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 700.860629] env[63175]: raise e [ 700.860629] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 700.860629] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 700.860629] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.860629] env[63175]: created_port_ids = self._update_ports_for_instance( [ 700.860629] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.860629] env[63175]: with excutils.save_and_reraise_exception(): [ 700.860629] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.860629] env[63175]: self.force_reraise() [ 700.860629] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.860629] env[63175]: raise self.value [ 700.860629] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.860629] env[63175]: updated_port = self._update_port( [ 700.860629] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.860629] env[63175]: _ensure_no_port_binding_failure(port) [ 700.860629] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.860629] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 700.861390] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 700.861390] env[63175]: Removing descriptor: 15 [ 700.893074] env[63175]: DEBUG nova.scheduler.client.report [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 701.012900] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 701.037043] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.037302] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.037457] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.037638] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.037783] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.038236] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.042081] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.042081] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.042081] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.042081] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.042081] env[63175]: DEBUG nova.virt.hardware [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.042243] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edeb84a4-4005-464e-a8fd-e744478ee4ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.048611] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbeb1a3-be25-4eb2-aa6c-3b10f7c365a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.063243] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Traceback (most recent call last): [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] yield resources [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.driver.spawn(context, instance, image_meta, [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] vm_ref = self.build_virtual_machine(instance, [ 701.063243] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] for vif in network_info: [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return self._sync_wrapper(fn, *args, **kwargs) [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.wait() [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self[:] = self._gt.wait() [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return self._exit_event.wait() [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.063805] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] current.throw(*self._exc) [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] result = function(*args, **kwargs) [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return func(*args, **kwargs) [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise e [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] nwinfo = self.network_api.allocate_for_instance( [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] created_port_ids = self._update_ports_for_instance( [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] with excutils.save_and_reraise_exception(): [ 701.064358] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.force_reraise() [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise self.value [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] updated_port = self._update_port( [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] _ensure_no_port_binding_failure(port) [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise exception.PortBindingFailed(port_id=port['id']) [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 701.064946] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] [ 701.064946] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Terminating instance [ 701.224108] env[63175]: DEBUG nova.network.neutron [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.286726] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.314344] env[63175]: DEBUG nova.network.neutron [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.398213] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.398765] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 701.401351] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.483s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.403184] env[63175]: INFO nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.567508] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.820639] env[63175]: DEBUG oslo_concurrency.lockutils [req-dfe16ace-bbb7-4a82-a320-07cd31f9b06a req-dfa9e7b1-dc1d-4470-8519-74e5553a8beb service nova] Releasing lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.820639] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.820639] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.908825] env[63175]: DEBUG nova.compute.utils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.912126] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 701.912305] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.972440] env[63175]: DEBUG nova.policy [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '816341e295cb4cdab052d643ad9a44d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ea6da3858c347d887dcfba2556fabad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.345079] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.395714] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Successfully created port: dba999cb-0c33-4bf9-be26-f2391c4990cd {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.413242] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 702.446816] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.725976] env[63175]: DEBUG nova.compute.manager [req-a2c2e186-d4aa-4675-8297-6d979f7a0f15 req-75edb313-4b7d-41c2-a292-b606b775985d service nova] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Received event network-vif-deleted-5e7ae472-9404-4627-8af8-2536c5af0daf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 702.804526] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8280a2c-1485-44ae-a2f1-d19fe92bdc54 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.810861] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce725ea8-189d-41f3-ba65-f1c67fc4dd2b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.840606] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8150ebec-a545-4e8e-a361-8ebcababc848 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.847785] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65813de-6b84-4995-bf62-a1ef2c4eb558 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.860746] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.948983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.949412] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 702.950496] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.950850] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b78a796-de3a-47ab-bf69-d4fdee9e42fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.959720] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cadddac-76ab-4d43-8fc2-1a026c07e3b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.984487] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b305bcb-87e0-4fc9-9579-56a5af9ace8e could not be found. [ 702.984872] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.985179] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 702.985879] env[63175]: DEBUG oslo.service.loopingcall [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.986220] env[63175]: DEBUG nova.compute.manager [-] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 702.986360] env[63175]: DEBUG nova.network.neutron [-] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.150529] env[63175]: DEBUG nova.network.neutron [-] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.364112] env[63175]: ERROR nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 703.364112] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.364112] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.364112] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.364112] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.364112] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.364112] env[63175]: ERROR nova.compute.manager raise self.value [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.364112] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.364112] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.364112] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.364618] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.364618] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.364618] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 703.364618] env[63175]: ERROR nova.compute.manager [ 703.364618] env[63175]: Traceback (most recent call last): [ 703.364618] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.364618] env[63175]: listener.cb(fileno) [ 703.364618] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.364618] env[63175]: result = function(*args, **kwargs) [ 703.364618] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.364618] env[63175]: return func(*args, **kwargs) [ 703.364618] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 703.364618] env[63175]: raise e [ 703.364618] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.364618] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 703.364618] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.364618] env[63175]: created_port_ids = self._update_ports_for_instance( [ 703.364618] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.364618] env[63175]: with excutils.save_and_reraise_exception(): [ 703.364618] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.364618] env[63175]: self.force_reraise() [ 703.364618] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.364618] env[63175]: raise self.value [ 703.364618] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.364618] env[63175]: updated_port = self._update_port( [ 703.364618] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.364618] env[63175]: _ensure_no_port_binding_failure(port) [ 703.364618] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.364618] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.365453] env[63175]: nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 703.365453] env[63175]: Removing descriptor: 15 [ 703.365453] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 703.425324] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 703.452214] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.452465] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.452622] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.452804] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.453155] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.453339] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.453546] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.453711] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.453862] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.454031] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.454206] env[63175]: DEBUG nova.virt.hardware [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.455317] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca434eb4-e00c-4a57-87f0-1c063efbe212 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.462799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9235140c-136c-4636-b0a6-b1c203685f7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.476146] env[63175]: ERROR nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Traceback (most recent call last): [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] yield resources [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.driver.spawn(context, instance, image_meta, [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] vm_ref = self.build_virtual_machine(instance, [ 703.476146] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] for vif in network_info: [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return self._sync_wrapper(fn, *args, **kwargs) [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.wait() [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self[:] = self._gt.wait() [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return self._exit_event.wait() [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.476725] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] current.throw(*self._exc) [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] result = function(*args, **kwargs) [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return func(*args, **kwargs) [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise e [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] nwinfo = self.network_api.allocate_for_instance( [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] created_port_ids = self._update_ports_for_instance( [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] with excutils.save_and_reraise_exception(): [ 703.477071] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.force_reraise() [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise self.value [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] updated_port = self._update_port( [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] _ensure_no_port_binding_failure(port) [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise exception.PortBindingFailed(port_id=port['id']) [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 703.477439] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] [ 703.477439] env[63175]: INFO nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Terminating instance [ 703.653491] env[63175]: DEBUG nova.network.neutron [-] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.870164] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.870827] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 703.873712] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.804s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.980758] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.980978] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.981208] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.156189] env[63175]: INFO nova.compute.manager [-] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Took 1.17 seconds to deallocate network for instance. [ 704.158561] env[63175]: DEBUG nova.compute.claims [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.158764] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.379075] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.383780] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 704.383908] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.441802] env[63175]: DEBUG nova.policy [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0a6c4c2d2594741ba8266842fc7289c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a20e7f44759460391b5390a70d619d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.503532] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.623475] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.742188] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab89738-31d4-46db-b642-a38c9132f4d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.750395] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e81f27-6ba5-4bfb-9dad-69d1df30b55b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.782026] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Successfully created port: b22182b2-5464-4f2e-af60-4118a04dff58 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.785130] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f554aa32-4a57-45be-bf25-0c43b51a7e5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.788322] env[63175]: DEBUG nova.compute.manager [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Received event network-changed-dba999cb-0c33-4bf9-be26-f2391c4990cd {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 704.788498] env[63175]: DEBUG nova.compute.manager [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Refreshing instance network info cache due to event network-changed-dba999cb-0c33-4bf9-be26-f2391c4990cd. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 704.788682] env[63175]: DEBUG oslo_concurrency.lockutils [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] Acquiring lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.795103] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b1a423-f592-4c72-9ef4-f195e98b0742 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.809800] env[63175]: DEBUG nova.compute.provider_tree [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.884585] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 705.125931] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.126395] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 705.126586] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.126922] env[63175]: DEBUG oslo_concurrency.lockutils [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] Acquired lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.127111] env[63175]: DEBUG nova.network.neutron [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Refreshing network info cache for port dba999cb-0c33-4bf9-be26-f2391c4990cd {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.128597] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9f22c88-3e48-4f4e-beb3-89bed5b3da74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.137627] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03311c97-c39d-4613-8ed3-527c98ba72b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.159982] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 182b1cc6-9626-4146-be9a-e15c5690aa1d could not be found. [ 705.160267] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.160450] env[63175]: INFO nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 705.160690] env[63175]: DEBUG oslo.service.loopingcall [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.160937] env[63175]: DEBUG nova.compute.manager [-] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 705.161045] env[63175]: DEBUG nova.network.neutron [-] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.190774] env[63175]: DEBUG nova.network.neutron [-] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.315015] env[63175]: DEBUG nova.scheduler.client.report [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 705.649888] env[63175]: DEBUG nova.network.neutron [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.690189] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 705.690189] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 705.690189] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.690189] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.690189] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.690189] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.690189] env[63175]: ERROR nova.compute.manager raise self.value [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.690189] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.690189] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.690189] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.690660] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.690660] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.690660] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 705.690660] env[63175]: ERROR nova.compute.manager [ 705.690660] env[63175]: Traceback (most recent call last): [ 705.690660] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.690660] env[63175]: listener.cb(fileno) [ 705.690660] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.690660] env[63175]: result = function(*args, **kwargs) [ 705.690660] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.690660] env[63175]: return func(*args, **kwargs) [ 705.690660] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 705.690660] env[63175]: raise e [ 705.690660] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 705.690660] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 705.690660] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.690660] env[63175]: created_port_ids = self._update_ports_for_instance( [ 705.690660] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.690660] env[63175]: with excutils.save_and_reraise_exception(): [ 705.690660] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.690660] env[63175]: self.force_reraise() [ 705.690660] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.690660] env[63175]: raise self.value [ 705.690660] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.690660] env[63175]: updated_port = self._update_port( [ 705.690660] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.690660] env[63175]: _ensure_no_port_binding_failure(port) [ 705.690660] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.690660] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.691466] env[63175]: nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 705.691466] env[63175]: Removing descriptor: 15 [ 705.693958] env[63175]: DEBUG nova.network.neutron [-] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.739139] env[63175]: DEBUG nova.network.neutron [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.819894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.820538] env[63175]: ERROR nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Traceback (most recent call last): [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.driver.spawn(context, instance, image_meta, [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] vm_ref = self.build_virtual_machine(instance, [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.820538] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] for vif in network_info: [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self._sync_wrapper(fn, *args, **kwargs) [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.wait() [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self[:] = self._gt.wait() [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self._exit_event.wait() [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] result = hub.switch() [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.820916] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return self.greenlet.switch() [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] result = function(*args, **kwargs) [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] return func(*args, **kwargs) [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise e [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] nwinfo = self.network_api.allocate_for_instance( [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] created_port_ids = self._update_ports_for_instance( [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] with excutils.save_and_reraise_exception(): [ 705.821250] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] self.force_reraise() [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise self.value [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] updated_port = self._update_port( [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] _ensure_no_port_binding_failure(port) [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] raise exception.PortBindingFailed(port_id=port['id']) [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] nova.exception.PortBindingFailed: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. [ 705.821563] env[63175]: ERROR nova.compute.manager [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] [ 705.821960] env[63175]: DEBUG nova.compute.utils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.822624] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.518s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.825758] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Build of instance 070514d0-ccc0-4fed-90bd-c7914760db09 was re-scheduled: Binding failed for port 04fc6cab-f27c-45da-bc49-68ec422110ea, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 705.826875] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 705.826875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquiring lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.826875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Acquired lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.826875] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.894045] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 705.917714] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.917956] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.918129] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.918318] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.918463] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.918637] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.918812] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.918969] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.919146] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.919304] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.919473] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.920332] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c409d9-4119-4650-8d4a-2e7d48a4c74e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.928219] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b2ce50-d590-482a-9266-459217baeaf1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.941890] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Traceback (most recent call last): [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] yield resources [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.driver.spawn(context, instance, image_meta, [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] vm_ref = self.build_virtual_machine(instance, [ 705.941890] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] for vif in network_info: [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return self._sync_wrapper(fn, *args, **kwargs) [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.wait() [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self[:] = self._gt.wait() [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return self._exit_event.wait() [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.942298] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] current.throw(*self._exc) [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] result = function(*args, **kwargs) [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return func(*args, **kwargs) [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise e [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] nwinfo = self.network_api.allocate_for_instance( [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] created_port_ids = self._update_ports_for_instance( [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] with excutils.save_and_reraise_exception(): [ 705.942689] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.force_reraise() [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise self.value [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] updated_port = self._update_port( [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] _ensure_no_port_binding_failure(port) [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise exception.PortBindingFailed(port_id=port['id']) [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 705.943102] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] [ 705.943102] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Terminating instance [ 706.196839] env[63175]: INFO nova.compute.manager [-] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Took 1.04 seconds to deallocate network for instance. [ 706.199219] env[63175]: DEBUG nova.compute.claims [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.199401] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.241497] env[63175]: DEBUG oslo_concurrency.lockutils [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] Releasing lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.241774] env[63175]: DEBUG nova.compute.manager [req-0fbe3cf4-1f8d-4faf-9828-fcea933e68c3 req-8f54f50e-adf4-404e-b779-8a459805eb0a service nova] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Received event network-vif-deleted-dba999cb-0c33-4bf9-be26-f2391c4990cd {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 706.345030] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.419883] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.446198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.446470] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.446543] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.622173] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c898559b-3ab7-4388-a7a7-da4d927f85fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.629477] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623891a7-963d-4dda-b7fe-19862336539f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.658267] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5bad8f-77b1-48e7-a2de-322cae63d5f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.665462] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e329200-f3c7-4b5f-bdd9-8ccde40f2263 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.679351] env[63175]: DEBUG nova.compute.provider_tree [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.794093] env[63175]: DEBUG nova.compute.manager [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Received event network-changed-b22182b2-5464-4f2e-af60-4118a04dff58 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 706.794248] env[63175]: DEBUG nova.compute.manager [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Refreshing instance network info cache due to event network-changed-b22182b2-5464-4f2e-af60-4118a04dff58. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 706.794432] env[63175]: DEBUG oslo_concurrency.lockutils [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] Acquiring lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.924458] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Releasing lock "refresh_cache-070514d0-ccc0-4fed-90bd-c7914760db09" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.924701] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 706.924881] env[63175]: DEBUG nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 706.925103] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.942070] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.962214] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.039033] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.183579] env[63175]: DEBUG nova.scheduler.client.report [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 707.444570] env[63175]: DEBUG nova.network.neutron [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.541712] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.542168] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 707.542360] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.542659] env[63175]: DEBUG oslo_concurrency.lockutils [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] Acquired lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.542826] env[63175]: DEBUG nova.network.neutron [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Refreshing network info cache for port b22182b2-5464-4f2e-af60-4118a04dff58 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.543856] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2d42543-672f-417a-ae87-e9fcac9ffd6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.553128] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9474f3a-6466-44f6-ad5d-32ded797ffca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.573883] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8aaf9c9f-e992-49ca-acad-d518503544e0 could not be found. [ 707.574089] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.574268] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 707.574495] env[63175]: DEBUG oslo.service.loopingcall [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.574701] env[63175]: DEBUG nova.compute.manager [-] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 707.574795] env[63175]: DEBUG nova.network.neutron [-] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.589305] env[63175]: DEBUG nova.network.neutron [-] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.688979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.689618] env[63175]: ERROR nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Traceback (most recent call last): [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.driver.spawn(context, instance, image_meta, [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] vm_ref = self.build_virtual_machine(instance, [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.689618] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] for vif in network_info: [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return self._sync_wrapper(fn, *args, **kwargs) [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.wait() [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self[:] = self._gt.wait() [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return self._exit_event.wait() [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] current.throw(*self._exc) [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.690192] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] result = function(*args, **kwargs) [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] return func(*args, **kwargs) [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise e [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] nwinfo = self.network_api.allocate_for_instance( [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] created_port_ids = self._update_ports_for_instance( [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] with excutils.save_and_reraise_exception(): [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] self.force_reraise() [ 707.690514] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise self.value [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] updated_port = self._update_port( [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] _ensure_no_port_binding_failure(port) [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] raise exception.PortBindingFailed(port_id=port['id']) [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] nova.exception.PortBindingFailed: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. [ 707.690890] env[63175]: ERROR nova.compute.manager [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] [ 707.690890] env[63175]: DEBUG nova.compute.utils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.691508] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.677s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.693341] env[63175]: INFO nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.696049] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Build of instance a5d426c1-9f30-43a7-942e-06cbda2fce30 was re-scheduled: Binding failed for port 761de51a-b058-4574-b5c7-512c29b1f8a2, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 707.696252] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 707.696473] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquiring lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.696620] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Acquired lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.696770] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.947803] env[63175]: INFO nova.compute.manager [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] [instance: 070514d0-ccc0-4fed-90bd-c7914760db09] Took 1.02 seconds to deallocate network for instance. [ 708.061551] env[63175]: DEBUG nova.network.neutron [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.091783] env[63175]: DEBUG nova.network.neutron [-] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.140500] env[63175]: DEBUG nova.network.neutron [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.214712] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.307761] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.595052] env[63175]: INFO nova.compute.manager [-] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Took 1.02 seconds to deallocate network for instance. [ 708.596311] env[63175]: DEBUG nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 708.596485] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.642422] env[63175]: DEBUG oslo_concurrency.lockutils [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] Releasing lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.642656] env[63175]: DEBUG nova.compute.manager [req-15814573-0201-4887-a40b-5ad967c2dbae req-fb072843-65af-416e-af95-826cd94056ae service nova] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Received event network-vif-deleted-b22182b2-5464-4f2e-af60-4118a04dff58 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 708.810986] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Releasing lock "refresh_cache-a5d426c1-9f30-43a7-942e-06cbda2fce30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.811254] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 708.811427] env[63175]: DEBUG nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 708.811590] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.834317] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.977503] env[63175]: INFO nova.scheduler.client.report [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Deleted allocations for instance 070514d0-ccc0-4fed-90bd-c7914760db09 [ 708.996849] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05eca7b1-f732-4cec-8d66-6bf45bbbd2e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.005362] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af40acd6-f0ae-42cc-bc3c-6f690a7c04b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.036198] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6673a71-36f8-4a4f-b693-20deec677ee5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.043602] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fe4c8a-5a2e-4ce6-b1e9-d42e8e6249e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.056364] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.338581] env[63175]: DEBUG nova.network.neutron [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.487105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6222d2c3-22e5-4d7e-83ad-92ea3f702056 tempest-ServerActionsTestOtherB-1214040972 tempest-ServerActionsTestOtherB-1214040972-project-member] Lock "070514d0-ccc0-4fed-90bd-c7914760db09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.437s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.558791] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 709.843694] env[63175]: INFO nova.compute.manager [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] [instance: a5d426c1-9f30-43a7-942e-06cbda2fce30] Took 1.03 seconds to deallocate network for instance. [ 709.989538] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 710.063360] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.064028] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 710.066569] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.318s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.515410] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.574598] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.577392] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 710.577534] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.620306] env[63175]: DEBUG nova.policy [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0a6c4c2d2594741ba8266842fc7289c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a20e7f44759460391b5390a70d619d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.887668] env[63175]: INFO nova.scheduler.client.report [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Deleted allocations for instance a5d426c1-9f30-43a7-942e-06cbda2fce30 [ 710.936131] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Successfully created port: 11a58963-ad08-4263-8bff-0c324253644d {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.957655] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f67479-970b-465f-b8f1-18dcba3812c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.966819] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3192ee0f-2215-4bbf-befc-0573e37a7d3c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.006374] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c52ff7-cfee-4b37-aef2-06514c2565cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.013979] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8377fcd3-67ca-4509-b608-5cb423330f41 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.026844] env[63175]: DEBUG nova.compute.provider_tree [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.081526] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 711.398183] env[63175]: DEBUG oslo_concurrency.lockutils [None req-04589261-3ebd-4422-ae6b-8159ef04fa59 tempest-ServerTagsTestJSON-309009601 tempest-ServerTagsTestJSON-309009601-project-member] Lock "a5d426c1-9f30-43a7-942e-06cbda2fce30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.509s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.529988] env[63175]: DEBUG nova.scheduler.client.report [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 711.757808] env[63175]: DEBUG nova.compute.manager [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Received event network-changed-11a58963-ad08-4263-8bff-0c324253644d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 711.757984] env[63175]: DEBUG nova.compute.manager [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Refreshing instance network info cache due to event network-changed-11a58963-ad08-4263-8bff-0c324253644d. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 711.758225] env[63175]: DEBUG oslo_concurrency.lockutils [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] Acquiring lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.758351] env[63175]: DEBUG oslo_concurrency.lockutils [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] Acquired lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.758822] env[63175]: DEBUG nova.network.neutron [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Refreshing network info cache for port 11a58963-ad08-4263-8bff-0c324253644d {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.901411] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 711.999539] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 711.999539] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.999539] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.999539] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.999539] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.999539] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.999539] env[63175]: ERROR nova.compute.manager raise self.value [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.999539] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.999539] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.999539] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.999986] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.999986] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.999986] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 711.999986] env[63175]: ERROR nova.compute.manager [ 711.999986] env[63175]: Traceback (most recent call last): [ 711.999986] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.999986] env[63175]: listener.cb(fileno) [ 711.999986] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.999986] env[63175]: result = function(*args, **kwargs) [ 711.999986] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.999986] env[63175]: return func(*args, **kwargs) [ 711.999986] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.999986] env[63175]: raise e [ 711.999986] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.999986] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 711.999986] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.999986] env[63175]: created_port_ids = self._update_ports_for_instance( [ 711.999986] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.999986] env[63175]: with excutils.save_and_reraise_exception(): [ 711.999986] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.999986] env[63175]: self.force_reraise() [ 711.999986] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.999986] env[63175]: raise self.value [ 711.999986] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.999986] env[63175]: updated_port = self._update_port( [ 711.999986] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.999986] env[63175]: _ensure_no_port_binding_failure(port) [ 711.999986] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.999986] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.000754] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 712.000754] env[63175]: Removing descriptor: 17 [ 712.035708] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.036485] env[63175]: ERROR nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Traceback (most recent call last): [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.driver.spawn(context, instance, image_meta, [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] vm_ref = self.build_virtual_machine(instance, [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.036485] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] for vif in network_info: [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return self._sync_wrapper(fn, *args, **kwargs) [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.wait() [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self[:] = self._gt.wait() [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return self._exit_event.wait() [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] current.throw(*self._exc) [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.036815] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] result = function(*args, **kwargs) [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] return func(*args, **kwargs) [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise e [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] nwinfo = self.network_api.allocate_for_instance( [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] created_port_ids = self._update_ports_for_instance( [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] with excutils.save_and_reraise_exception(): [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] self.force_reraise() [ 712.037303] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise self.value [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] updated_port = self._update_port( [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] _ensure_no_port_binding_failure(port) [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] raise exception.PortBindingFailed(port_id=port['id']) [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] nova.exception.PortBindingFailed: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. [ 712.037615] env[63175]: ERROR nova.compute.manager [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] [ 712.037615] env[63175]: DEBUG nova.compute.utils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.038654] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.801s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.041522] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Build of instance 7a25b67c-47b8-420f-9aa2-8c296b8ab221 was re-scheduled: Binding failed for port 08649e21-1bb9-4b4c-9b42-5254e44d28c2, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 712.043237] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 712.043237] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.043237] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.043237] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.097435] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 712.121570] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.121827] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.122673] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.122673] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.122673] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.122673] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.122673] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.122883] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.123242] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.123242] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.123369] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.124256] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6abc92-1888-4f1d-85a4-0d6e2ebd06be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.132710] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f38ce06-310d-4b65-baa4-1d1436f256db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.147388] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Traceback (most recent call last): [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] yield resources [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.driver.spawn(context, instance, image_meta, [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] vm_ref = self.build_virtual_machine(instance, [ 712.147388] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] for vif in network_info: [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return self._sync_wrapper(fn, *args, **kwargs) [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.wait() [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self[:] = self._gt.wait() [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return self._exit_event.wait() [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.147857] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] current.throw(*self._exc) [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] result = function(*args, **kwargs) [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return func(*args, **kwargs) [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise e [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] nwinfo = self.network_api.allocate_for_instance( [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] created_port_ids = self._update_ports_for_instance( [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] with excutils.save_and_reraise_exception(): [ 712.148303] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.force_reraise() [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise self.value [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] updated_port = self._update_port( [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] _ensure_no_port_binding_failure(port) [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise exception.PortBindingFailed(port_id=port['id']) [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 712.148745] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] [ 712.148745] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Terminating instance [ 712.284187] env[63175]: DEBUG nova.network.neutron [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.428461] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.507459] env[63175]: DEBUG nova.network.neutron [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.572911] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.656387] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.789355] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.904335] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8265ef8b-e09a-47af-92c1-2d87e25ead3f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.911598] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2e5e3f-aed1-40e5-af2f-fea5cac64816 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.942938] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7912f9-f193-4af2-b583-28b93c988886 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.950387] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70a9f35-443e-4632-9d50-56448dc268f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.963244] env[63175]: DEBUG nova.compute.provider_tree [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.009714] env[63175]: DEBUG oslo_concurrency.lockutils [req-141a8154-70c9-4441-9120-14a5d2d237e8 req-bfd01710-c906-43e2-add2-cbe226be712f service nova] Releasing lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.010139] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.010324] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.293709] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Releasing lock "refresh_cache-7a25b67c-47b8-420f-9aa2-8c296b8ab221" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.293935] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 713.294127] env[63175]: DEBUG nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 713.294290] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.342798] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.465959] env[63175]: DEBUG nova.scheduler.client.report [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 713.542136] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.715538] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.810852] env[63175]: DEBUG nova.compute.manager [req-3cc9eff2-f6a1-48a8-86a9-3b20d98d794c req-167c6d38-ea28-41fb-9195-1f0ed1e5580f service nova] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Received event network-vif-deleted-11a58963-ad08-4263-8bff-0c324253644d {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 713.846300] env[63175]: DEBUG nova.network.neutron [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.975021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.975021] env[63175]: ERROR nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Traceback (most recent call last): [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.driver.spawn(context, instance, image_meta, [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.975021] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] vm_ref = self.build_virtual_machine(instance, [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] for vif in network_info: [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self._sync_wrapper(fn, *args, **kwargs) [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.wait() [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self[:] = self._gt.wait() [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self._exit_event.wait() [ 713.975565] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] result = hub.switch() [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return self.greenlet.switch() [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] result = function(*args, **kwargs) [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] return func(*args, **kwargs) [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise e [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] nwinfo = self.network_api.allocate_for_instance( [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.975941] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] created_port_ids = self._update_ports_for_instance( [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] with excutils.save_and_reraise_exception(): [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] self.force_reraise() [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise self.value [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] updated_port = self._update_port( [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] _ensure_no_port_binding_failure(port) [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.976327] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] raise exception.PortBindingFailed(port_id=port['id']) [ 713.976637] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] nova.exception.PortBindingFailed: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. [ 713.976637] env[63175]: ERROR nova.compute.manager [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] [ 713.976637] env[63175]: DEBUG nova.compute.utils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.976637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.600s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.979695] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Build of instance d85e98e9-d3d5-4781-8428-d8ab517be146 was re-scheduled: Binding failed for port 43932b33-1509-4d42-8684-f7093e2aebc9, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 713.979840] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 713.980116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquiring lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.983019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Acquired lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.983019] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.221320] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.221741] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 714.225023] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.225023] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c401acd-769c-4c8e-817a-d34d40abe60b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.236103] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2b6750-f2f0-4f83-95e5-bfa39cb1270b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.261264] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 223ccdbc-cee3-4040-a1ee-a53929025926 could not be found. [ 714.261512] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.262264] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Took 0.04 seconds to destroy the instance on the hypervisor. [ 714.262583] env[63175]: DEBUG oslo.service.loopingcall [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.262802] env[63175]: DEBUG nova.compute.manager [-] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 714.262899] env[63175]: DEBUG nova.network.neutron [-] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.285930] env[63175]: DEBUG nova.network.neutron [-] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.354532] env[63175]: INFO nova.compute.manager [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 7a25b67c-47b8-420f-9aa2-8c296b8ab221] Took 1.06 seconds to deallocate network for instance. [ 714.498919] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.614266] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.789734] env[63175]: DEBUG nova.network.neutron [-] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.847936] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde4a2fc-91f3-4923-a400-397be4eaddb9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.866198] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fff21a-71d4-4327-b685-9a0a6b5cf82c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.898873] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c26f399-766c-46b0-8272-0c0da2d1e148 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.908510] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a256aaa-c3a8-4fba-b08e-4bce088c8890 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.924850] env[63175]: DEBUG nova.compute.provider_tree [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.116598] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Releasing lock "refresh_cache-d85e98e9-d3d5-4781-8428-d8ab517be146" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.116816] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 715.117396] env[63175]: DEBUG nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 715.117396] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.135042] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.292768] env[63175]: INFO nova.compute.manager [-] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Took 1.03 seconds to deallocate network for instance. [ 715.297303] env[63175]: DEBUG nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.297303] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.414905] env[63175]: INFO nova.scheduler.client.report [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Deleted allocations for instance 7a25b67c-47b8-420f-9aa2-8c296b8ab221 [ 715.427692] env[63175]: DEBUG nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 715.638220] env[63175]: DEBUG nova.network.neutron [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.924515] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e856608e-eb48-4bed-b237-1676c7f556d1 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "7a25b67c-47b8-420f-9aa2-8c296b8ab221" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.625s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.938169] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.938844] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Traceback (most recent call last): [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.driver.spawn(context, instance, image_meta, [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] vm_ref = self.build_virtual_machine(instance, [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.938844] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] for vif in network_info: [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return self._sync_wrapper(fn, *args, **kwargs) [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.wait() [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self[:] = self._gt.wait() [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return self._exit_event.wait() [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] current.throw(*self._exc) [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.939203] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] result = function(*args, **kwargs) [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] return func(*args, **kwargs) [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise e [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] nwinfo = self.network_api.allocate_for_instance( [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] created_port_ids = self._update_ports_for_instance( [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] with excutils.save_and_reraise_exception(): [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] self.force_reraise() [ 715.939515] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise self.value [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] updated_port = self._update_port( [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] _ensure_no_port_binding_failure(port) [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] raise exception.PortBindingFailed(port_id=port['id']) [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] nova.exception.PortBindingFailed: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. [ 715.939887] env[63175]: ERROR nova.compute.manager [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] [ 715.939887] env[63175]: DEBUG nova.compute.utils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.941414] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.655s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.942799] env[63175]: INFO nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.946206] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Build of instance 11747695-c79c-477d-b5ae-44c49dfb4bba was re-scheduled: Binding failed for port de487c35-048b-4d84-9446-952e74b58d7b, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 715.946671] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 715.946890] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.947042] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.947195] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.951292] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "72d131a0-2617-49a1-8aff-897908929bb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.951496] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.147235] env[63175]: INFO nova.compute.manager [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] [instance: d85e98e9-d3d5-4781-8428-d8ab517be146] Took 1.03 seconds to deallocate network for instance. [ 716.427205] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 716.486539] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.564902] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.857959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "4d07a537-7df6-4659-8760-bf7e7925da25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.858170] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.950535] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.069278] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-11747695-c79c-477d-b5ae-44c49dfb4bba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.069408] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 717.069552] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 717.069694] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.086500] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.172141] env[63175]: INFO nova.scheduler.client.report [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Deleted allocations for instance d85e98e9-d3d5-4781-8428-d8ab517be146 [ 717.233398] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecba50ed-c0af-4361-a4ee-1413f3bbb2f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.239888] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b17186a-7e1b-4519-b626-aa82cae4cda6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.270009] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1e1239-963a-4c91-b625-c366f6f34b4f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.277061] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8682cde2-4b78-457b-a396-3884f1f13e1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.289538] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.589413] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.680843] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4f5fc93-41d0-45c0-9fd3-141fea95cc60 tempest-ServerMetadataTestJSON-294456322 tempest-ServerMetadataTestJSON-294456322-project-member] Lock "d85e98e9-d3d5-4781-8428-d8ab517be146" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.891s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.793757] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 718.094998] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 11747695-c79c-477d-b5ae-44c49dfb4bba] Took 1.02 seconds to deallocate network for instance. [ 718.184028] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 718.300514] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.300514] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 718.301928] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.143s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.707195] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.806395] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.811013] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 718.811195] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.874295] env[63175]: DEBUG nova.policy [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0a6c4c2d2594741ba8266842fc7289c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a20e7f44759460391b5390a70d619d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.111081] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff59fb0a-8186-414a-9fe9-cc3282d7abd5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.122690] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f67a626-face-4576-a1e3-a5b3bed1bd53 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.156164] env[63175]: INFO nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted allocations for instance 11747695-c79c-477d-b5ae-44c49dfb4bba [ 719.162294] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d97c438-5ea8-4fcc-8180-6a25e6c5e633 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.171684] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72f616a-4088-4022-a080-2c8480dacf74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.185683] env[63175]: DEBUG nova.compute.provider_tree [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.314213] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.481756] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Successfully created port: 305905a8-3ff7-452e-b303-c6177f0605d6 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.670469] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "11747695-c79c-477d-b5ae-44c49dfb4bba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.512s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.688760] env[63175]: DEBUG nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 720.174032] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 720.196441] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.196441] env[63175]: ERROR nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Traceback (most recent call last): [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.driver.spawn(context, instance, image_meta, [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.196441] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] vm_ref = self.build_virtual_machine(instance, [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] for vif in network_info: [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return self._sync_wrapper(fn, *args, **kwargs) [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.wait() [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self[:] = self._gt.wait() [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return self._exit_event.wait() [ 720.196925] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] current.throw(*self._exc) [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] result = function(*args, **kwargs) [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] return func(*args, **kwargs) [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise e [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] nwinfo = self.network_api.allocate_for_instance( [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] created_port_ids = self._update_ports_for_instance( [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.197652] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] with excutils.save_and_reraise_exception(): [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] self.force_reraise() [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise self.value [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] updated_port = self._update_port( [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] _ensure_no_port_binding_failure(port) [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] raise exception.PortBindingFailed(port_id=port['id']) [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] nova.exception.PortBindingFailed: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. [ 720.198299] env[63175]: ERROR nova.compute.manager [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] [ 720.198791] env[63175]: DEBUG nova.compute.utils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.198791] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.998s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.200795] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Build of instance 8b305bcb-87e0-4fc9-9579-56a5af9ace8e was re-scheduled: Binding failed for port 5e7ae472-9404-4627-8af8-2536c5af0daf, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 720.201651] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 720.201827] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.202022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.202203] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.235084] env[63175]: DEBUG nova.compute.manager [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Received event network-changed-305905a8-3ff7-452e-b303-c6177f0605d6 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 720.235214] env[63175]: DEBUG nova.compute.manager [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Refreshing instance network info cache due to event network-changed-305905a8-3ff7-452e-b303-c6177f0605d6. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 720.235442] env[63175]: DEBUG oslo_concurrency.lockutils [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] Acquiring lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.235590] env[63175]: DEBUG oslo_concurrency.lockutils [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] Acquired lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.235751] env[63175]: DEBUG nova.network.neutron [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Refreshing network info cache for port 305905a8-3ff7-452e-b303-c6177f0605d6 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.321363] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 720.354263] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.354531] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.354668] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.354848] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.354992] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.355168] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.355373] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.355531] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.355696] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.355854] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.356032] env[63175]: DEBUG nova.virt.hardware [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.356871] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369b07c3-7d85-4cb3-a290-723c2761951a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.366354] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd37b546-0ee9-4c3d-89f5-bbb140ae067c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.463550] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 720.463550] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.463550] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.463550] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.463550] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.463550] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.463550] env[63175]: ERROR nova.compute.manager raise self.value [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.463550] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.463550] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.463550] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.464047] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.464047] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.464047] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 720.464047] env[63175]: ERROR nova.compute.manager [ 720.464047] env[63175]: Traceback (most recent call last): [ 720.464047] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.464047] env[63175]: listener.cb(fileno) [ 720.464047] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.464047] env[63175]: result = function(*args, **kwargs) [ 720.464047] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.464047] env[63175]: return func(*args, **kwargs) [ 720.464047] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.464047] env[63175]: raise e [ 720.464047] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.464047] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 720.464047] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.464047] env[63175]: created_port_ids = self._update_ports_for_instance( [ 720.464047] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.464047] env[63175]: with excutils.save_and_reraise_exception(): [ 720.464047] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.464047] env[63175]: self.force_reraise() [ 720.464047] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.464047] env[63175]: raise self.value [ 720.464047] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.464047] env[63175]: updated_port = self._update_port( [ 720.464047] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.464047] env[63175]: _ensure_no_port_binding_failure(port) [ 720.464047] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.464047] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.464824] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 720.464824] env[63175]: Removing descriptor: 17 [ 720.464824] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Traceback (most recent call last): [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] yield resources [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.driver.spawn(context, instance, image_meta, [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.464824] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] vm_ref = self.build_virtual_machine(instance, [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] for vif in network_info: [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self._sync_wrapper(fn, *args, **kwargs) [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.wait() [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self[:] = self._gt.wait() [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self._exit_event.wait() [ 720.465184] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] result = hub.switch() [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self.greenlet.switch() [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] result = function(*args, **kwargs) [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return func(*args, **kwargs) [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise e [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] nwinfo = self.network_api.allocate_for_instance( [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.465554] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] created_port_ids = self._update_ports_for_instance( [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] with excutils.save_and_reraise_exception(): [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.force_reraise() [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise self.value [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] updated_port = self._update_port( [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] _ensure_no_port_binding_failure(port) [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.465940] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise exception.PortBindingFailed(port_id=port['id']) [ 720.466285] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 720.466285] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] [ 720.466285] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Terminating instance [ 720.705247] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.723554] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.764172] env[63175]: DEBUG nova.network.neutron [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.828610] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.854669] env[63175]: DEBUG nova.network.neutron [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.969134] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.995513] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5c494b-5d34-4b82-8514-9560da175b80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.003734] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d56118-fded-4315-bc13-506aa2a0b44b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.035664] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d4f8f2-81ab-4dd2-8fb9-a7aad4296fbc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.043858] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf27637-2fa8-4f55-83aa-5e5e343263ef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.059470] env[63175]: DEBUG nova.compute.provider_tree [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.333629] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-8b305bcb-87e0-4fc9-9579-56a5af9ace8e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.333882] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 721.334073] env[63175]: DEBUG nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 721.334247] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.352962] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.360511] env[63175]: DEBUG oslo_concurrency.lockutils [req-bea01a58-fbd5-44bc-bfb4-6b877b38540f req-9d57b446-523b-459e-9f04-5897050957ab service nova] Releasing lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.360948] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.361105] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.564173] env[63175]: DEBUG nova.scheduler.client.report [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 721.859582] env[63175]: DEBUG nova.network.neutron [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.889442] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.971573] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.071806] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.072404] env[63175]: ERROR nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Traceback (most recent call last): [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.driver.spawn(context, instance, image_meta, [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] vm_ref = self.build_virtual_machine(instance, [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.072404] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] for vif in network_info: [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return self._sync_wrapper(fn, *args, **kwargs) [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.wait() [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self[:] = self._gt.wait() [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return self._exit_event.wait() [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] current.throw(*self._exc) [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.072791] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] result = function(*args, **kwargs) [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] return func(*args, **kwargs) [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise e [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] nwinfo = self.network_api.allocate_for_instance( [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] created_port_ids = self._update_ports_for_instance( [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] with excutils.save_and_reraise_exception(): [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] self.force_reraise() [ 722.073202] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise self.value [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] updated_port = self._update_port( [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] _ensure_no_port_binding_failure(port) [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] raise exception.PortBindingFailed(port_id=port['id']) [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] nova.exception.PortBindingFailed: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. [ 722.073545] env[63175]: ERROR nova.compute.manager [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] [ 722.073545] env[63175]: DEBUG nova.compute.utils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.074712] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.478s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.084919] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Build of instance 182b1cc6-9626-4146-be9a-e15c5690aa1d was re-scheduled: Binding failed for port dba999cb-0c33-4bf9-be26-f2391c4990cd, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 722.085272] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 722.085503] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.085649] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.085807] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.265575] env[63175]: DEBUG nova.compute.manager [req-e2551dfb-0b81-4553-8e52-021b0467cab3 req-1d35dd8d-ea4a-4399-b834-ebede8850826 service nova] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Received event network-vif-deleted-305905a8-3ff7-452e-b303-c6177f0605d6 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 722.364184] env[63175]: INFO nova.compute.manager [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 8b305bcb-87e0-4fc9-9579-56a5af9ace8e] Took 1.03 seconds to deallocate network for instance. [ 722.474611] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.475051] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 722.475822] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.475822] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54163291-9df8-49ec-9520-87d6aef06ed1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.484709] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc1d018-ebc1-4541-80b3-d08715d449fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.509239] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5bc39a7-a493-41a3-a2f4-8958883adca8 could not be found. [ 722.509365] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.509537] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 722.509788] env[63175]: DEBUG oslo.service.loopingcall [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.510102] env[63175]: DEBUG nova.compute.manager [-] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 722.510163] env[63175]: DEBUG nova.network.neutron [-] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.529931] env[63175]: DEBUG nova.network.neutron [-] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.606639] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.729842] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.981712] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e74f3d-90ab-4cd2-b00f-17e3799e60bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.989278] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d5bc92-30f3-47b3-88de-eaa1e0a84eb0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.018929] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af30c239-3657-42f8-aa3a-08b7929d18bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.026285] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20c18bb-8241-4b4f-b330-0f7269e282d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.039139] env[63175]: DEBUG nova.network.neutron [-] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.042861] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.233018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "refresh_cache-182b1cc6-9626-4146-be9a-e15c5690aa1d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.233298] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 723.233509] env[63175]: DEBUG nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 723.233680] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.252592] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.398948] env[63175]: INFO nova.scheduler.client.report [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted allocations for instance 8b305bcb-87e0-4fc9-9579-56a5af9ace8e [ 723.544309] env[63175]: INFO nova.compute.manager [-] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Took 1.03 seconds to deallocate network for instance. [ 723.548796] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 723.556189] env[63175]: DEBUG nova.compute.claims [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.556189] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.756301] env[63175]: DEBUG nova.network.neutron [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.906911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0b24b086-c880-456b-9f31-cc02764da74e tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "8b305bcb-87e0-4fc9-9579-56a5af9ace8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.703s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.056740] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.058060] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Traceback (most recent call last): [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.driver.spawn(context, instance, image_meta, [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] vm_ref = self.build_virtual_machine(instance, [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.058060] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] for vif in network_info: [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return self._sync_wrapper(fn, *args, **kwargs) [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.wait() [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self[:] = self._gt.wait() [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return self._exit_event.wait() [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] current.throw(*self._exc) [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.058471] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] result = function(*args, **kwargs) [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] return func(*args, **kwargs) [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise e [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] nwinfo = self.network_api.allocate_for_instance( [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] created_port_ids = self._update_ports_for_instance( [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] with excutils.save_and_reraise_exception(): [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] self.force_reraise() [ 724.058868] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise self.value [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] updated_port = self._update_port( [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] _ensure_no_port_binding_failure(port) [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] raise exception.PortBindingFailed(port_id=port['id']) [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] nova.exception.PortBindingFailed: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. [ 724.059304] env[63175]: ERROR nova.compute.manager [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] [ 724.059304] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.061151] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.545s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.063439] env[63175]: INFO nova.compute.claims [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.066522] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Build of instance 8aaf9c9f-e992-49ca-acad-d518503544e0 was re-scheduled: Binding failed for port b22182b2-5464-4f2e-af60-4118a04dff58, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 724.066976] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 724.067259] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.067405] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.067563] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.258213] env[63175]: INFO nova.compute.manager [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 182b1cc6-9626-4146-be9a-e15c5690aa1d] Took 1.02 seconds to deallocate network for instance. [ 724.386597] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.386864] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.412361] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 724.589888] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.707809] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.935934] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.211253] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-8aaf9c9f-e992-49ca-acad-d518503544e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.211517] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 725.211705] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 725.211876] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.228211] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.284155] env[63175]: INFO nova.scheduler.client.report [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted allocations for instance 182b1cc6-9626-4146-be9a-e15c5690aa1d [ 725.344701] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb10645-3391-40df-975c-6a66e3d6c4b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.352324] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a4c298-44fb-46b4-bc0e-25231696fe6f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.382562] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb2168c-1f8b-4733-894f-46a753b27cbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.389834] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346e3bd2-2345-4d66-a09e-fe51ac8850c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.402874] env[63175]: DEBUG nova.compute.provider_tree [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.481902] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "780989ee-98a1-4e99-9014-45de6921d4bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.482199] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.504352] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.504569] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.731134] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.793859] env[63175]: DEBUG oslo_concurrency.lockutils [None req-84a938b4-cf0a-4f68-b3a7-2c8c60d9f3ee tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "182b1cc6-9626-4146-be9a-e15c5690aa1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.993s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.908406] env[63175]: DEBUG nova.scheduler.client.report [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 726.236028] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 8aaf9c9f-e992-49ca-acad-d518503544e0] Took 1.02 seconds to deallocate network for instance. [ 726.296914] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 726.413583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.413982] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 726.416430] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.988s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.418205] env[63175]: INFO nova.compute.claims [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.818074] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.924190] env[63175]: DEBUG nova.compute.utils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.931415] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 726.931627] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.991298] env[63175]: DEBUG nova.policy [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e0bf05d334444e4ca1d35ae10c972ec7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8e558fa473141a68a95719bec89b2f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.262509] env[63175]: INFO nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Deleted allocations for instance 8aaf9c9f-e992-49ca-acad-d518503544e0 [ 727.350916] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Successfully created port: dd1e9fb1-3cbf-4653-a505-c630f36c5e9e {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.432851] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 727.774086] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "8aaf9c9f-e992-49ca-acad-d518503544e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.356s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.816982] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84475d8-a453-40cb-a64d-e9b15bd62228 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.825777] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c24d6fd-c4ef-4b43-b77e-66593e429454 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.856622] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882e7596-c444-4dac-8e05-de1cb42755db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.865096] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6460f4-e749-494e-8b80-4c9f719151b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.880361] env[63175]: DEBUG nova.compute.provider_tree [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.278613] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 728.285157] env[63175]: DEBUG nova.compute.manager [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Received event network-changed-dd1e9fb1-3cbf-4653-a505-c630f36c5e9e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 728.285157] env[63175]: DEBUG nova.compute.manager [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Refreshing instance network info cache due to event network-changed-dd1e9fb1-3cbf-4653-a505-c630f36c5e9e. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 728.285157] env[63175]: DEBUG oslo_concurrency.lockutils [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] Acquiring lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.285157] env[63175]: DEBUG oslo_concurrency.lockutils [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] Acquired lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.285157] env[63175]: DEBUG nova.network.neutron [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Refreshing network info cache for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.384096] env[63175]: DEBUG nova.scheduler.client.report [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 728.438439] env[63175]: ERROR nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 728.438439] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.438439] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.438439] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.438439] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.438439] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.438439] env[63175]: ERROR nova.compute.manager raise self.value [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.438439] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.438439] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.438439] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.438987] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.438987] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.438987] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 728.438987] env[63175]: ERROR nova.compute.manager [ 728.438987] env[63175]: Traceback (most recent call last): [ 728.438987] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.438987] env[63175]: listener.cb(fileno) [ 728.438987] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.438987] env[63175]: result = function(*args, **kwargs) [ 728.438987] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.438987] env[63175]: return func(*args, **kwargs) [ 728.438987] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 728.438987] env[63175]: raise e [ 728.438987] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.438987] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 728.438987] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.438987] env[63175]: created_port_ids = self._update_ports_for_instance( [ 728.438987] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.438987] env[63175]: with excutils.save_and_reraise_exception(): [ 728.438987] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.438987] env[63175]: self.force_reraise() [ 728.438987] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.438987] env[63175]: raise self.value [ 728.438987] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.438987] env[63175]: updated_port = self._update_port( [ 728.438987] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.438987] env[63175]: _ensure_no_port_binding_failure(port) [ 728.438987] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.438987] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.439931] env[63175]: nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 728.439931] env[63175]: Removing descriptor: 15 [ 728.446711] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 728.467231] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:45:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b472f8e6-0c9b-495d-95d3-f3787c4ff99c',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-382974288',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.467478] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.467639] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.467868] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.468032] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.468190] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.468399] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.468558] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.468783] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.468949] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.469220] env[63175]: DEBUG nova.virt.hardware [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.470166] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efe1391-d4d2-4228-86bf-2ba868805e52 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.478560] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f52dc3c-349b-49a7-924d-e38f70e3746d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.492604] env[63175]: ERROR nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Traceback (most recent call last): [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] yield resources [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.driver.spawn(context, instance, image_meta, [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] vm_ref = self.build_virtual_machine(instance, [ 728.492604] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] for vif in network_info: [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return self._sync_wrapper(fn, *args, **kwargs) [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.wait() [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self[:] = self._gt.wait() [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return self._exit_event.wait() [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.493056] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] current.throw(*self._exc) [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] result = function(*args, **kwargs) [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return func(*args, **kwargs) [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise e [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] nwinfo = self.network_api.allocate_for_instance( [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] created_port_ids = self._update_ports_for_instance( [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] with excutils.save_and_reraise_exception(): [ 728.493498] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.force_reraise() [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise self.value [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] updated_port = self._update_port( [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] _ensure_no_port_binding_failure(port) [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise exception.PortBindingFailed(port_id=port['id']) [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 728.494052] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] [ 728.494052] env[63175]: INFO nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Terminating instance [ 728.806840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.807894] env[63175]: DEBUG nova.network.neutron [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.866394] env[63175]: DEBUG nova.network.neutron [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.871536] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.871758] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.889304] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.889755] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 728.892341] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.597s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.997704] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.372713] env[63175]: DEBUG oslo_concurrency.lockutils [req-11157d4b-fa7f-4ef7-85d2-510917cb12eb req-2e6cc219-742e-4d0e-8e73-095e6f619437 service nova] Releasing lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.373202] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquired lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.373433] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.396885] env[63175]: DEBUG nova.compute.utils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.401647] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 729.401647] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.457428] env[63175]: DEBUG nova.policy [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb32cec66aac454abeac265019302c52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56ec9e0495f5445e943bd85cca84cc5e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.684165] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3869b706-cfb4-40a7-ae09-1a5214b33704 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.691677] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c496e6b9-604a-40e8-8aa3-e280c912c6bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.721293] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9ee5b6-b82f-4155-8d19-f9fcb1838175 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.728978] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e942fb-6a2c-41c7-80a3-e802ea682101 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.742430] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.784904] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Successfully created port: bab7f3c1-eee7-4a3c-985a-91ec5b148243 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.895651] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.901561] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 729.963574] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.247145] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 730.312627] env[63175]: DEBUG nova.compute.manager [req-9e687291-407e-4b85-a575-23acc658d632 req-e3fe9f6d-2f59-471c-ba0a-11b2278cd002 service nova] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Received event network-vif-deleted-dd1e9fb1-3cbf-4653-a505-c630f36c5e9e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 730.468190] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Releasing lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.468495] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 730.468692] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.469080] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-793b15cf-a5cb-42ba-8397-68224cf4e358 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.479652] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3907e80-c738-4a8d-85c1-e486ef0e1356 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.503088] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9af5d212-c032-4446-b3cd-87fe4e66ad7c could not be found. [ 730.503388] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.503577] env[63175]: INFO nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 730.503807] env[63175]: DEBUG oslo.service.loopingcall [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.503999] env[63175]: DEBUG nova.compute.manager [-] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 730.504101] env[63175]: DEBUG nova.network.neutron [-] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.525977] env[63175]: DEBUG nova.network.neutron [-] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.691605] env[63175]: ERROR nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 730.691605] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.691605] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.691605] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.691605] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.691605] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.691605] env[63175]: ERROR nova.compute.manager raise self.value [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.691605] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.691605] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.691605] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.692169] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.692169] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.692169] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 730.692169] env[63175]: ERROR nova.compute.manager [ 730.692169] env[63175]: Traceback (most recent call last): [ 730.692169] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.692169] env[63175]: listener.cb(fileno) [ 730.692169] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.692169] env[63175]: result = function(*args, **kwargs) [ 730.692169] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.692169] env[63175]: return func(*args, **kwargs) [ 730.692169] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.692169] env[63175]: raise e [ 730.692169] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.692169] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 730.692169] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.692169] env[63175]: created_port_ids = self._update_ports_for_instance( [ 730.692169] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.692169] env[63175]: with excutils.save_and_reraise_exception(): [ 730.692169] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.692169] env[63175]: self.force_reraise() [ 730.692169] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.692169] env[63175]: raise self.value [ 730.692169] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.692169] env[63175]: updated_port = self._update_port( [ 730.692169] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.692169] env[63175]: _ensure_no_port_binding_failure(port) [ 730.692169] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.692169] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.693108] env[63175]: nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 730.693108] env[63175]: Removing descriptor: 15 [ 730.753343] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.753989] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Traceback (most recent call last): [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.driver.spawn(context, instance, image_meta, [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] vm_ref = self.build_virtual_machine(instance, [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.753989] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] for vif in network_info: [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return self._sync_wrapper(fn, *args, **kwargs) [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.wait() [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self[:] = self._gt.wait() [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return self._exit_event.wait() [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] current.throw(*self._exc) [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.754438] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] result = function(*args, **kwargs) [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] return func(*args, **kwargs) [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise e [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] nwinfo = self.network_api.allocate_for_instance( [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] created_port_ids = self._update_ports_for_instance( [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] with excutils.save_and_reraise_exception(): [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] self.force_reraise() [ 730.754839] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise self.value [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] updated_port = self._update_port( [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] _ensure_no_port_binding_failure(port) [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] raise exception.PortBindingFailed(port_id=port['id']) [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] nova.exception.PortBindingFailed: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. [ 730.755269] env[63175]: ERROR nova.compute.manager [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] [ 730.755269] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.755854] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.805s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.757337] env[63175]: INFO nova.compute.claims [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.760019] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Build of instance 223ccdbc-cee3-4040-a1ee-a53929025926 was re-scheduled: Binding failed for port 11a58963-ad08-4263-8bff-0c324253644d, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 730.760432] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 730.760652] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.760797] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.760952] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.913980] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 730.939048] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.939328] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.939484] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.939666] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.939812] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.939956] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.940178] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.940336] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.940495] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.940650] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.940826] env[63175]: DEBUG nova.virt.hardware [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.941741] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93eaacd3-7d94-4bad-9595-b77e04e75603 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.950700] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec4edac-9508-4565-b324-a84acf9963f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.963753] env[63175]: ERROR nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] Traceback (most recent call last): [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] yield resources [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.driver.spawn(context, instance, image_meta, [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] vm_ref = self.build_virtual_machine(instance, [ 730.963753] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] for vif in network_info: [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return self._sync_wrapper(fn, *args, **kwargs) [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.wait() [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self[:] = self._gt.wait() [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return self._exit_event.wait() [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.964188] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] current.throw(*self._exc) [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] result = function(*args, **kwargs) [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return func(*args, **kwargs) [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise e [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] nwinfo = self.network_api.allocate_for_instance( [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] created_port_ids = self._update_ports_for_instance( [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] with excutils.save_and_reraise_exception(): [ 730.964661] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.force_reraise() [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise self.value [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] updated_port = self._update_port( [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] _ensure_no_port_binding_failure(port) [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise exception.PortBindingFailed(port_id=port['id']) [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 730.965117] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] [ 730.965117] env[63175]: INFO nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Terminating instance [ 731.028824] env[63175]: DEBUG nova.network.neutron [-] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.281554] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.377823] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.469496] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquiring lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.469681] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquired lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.469860] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.532740] env[63175]: INFO nova.compute.manager [-] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Took 1.03 seconds to deallocate network for instance. [ 731.535699] env[63175]: DEBUG nova.compute.claims [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.535850] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.881848] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-223ccdbc-cee3-4040-a1ee-a53929025926" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.882132] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 731.882332] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 731.882502] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.902962] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.990600] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.041094] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff04e766-34a6-4cf4-ad7b-4f885825e4fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.048924] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f60895-d823-42b9-802f-e6586b0fe4a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.079081] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.081139] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2d031e-1c21-4e32-9a9b-b163884b2879 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.088199] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b527c17-6164-434e-8a95-3dcb448ec89d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.101954] env[63175]: DEBUG nova.compute.provider_tree [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.337287] env[63175]: DEBUG nova.compute.manager [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Received event network-changed-bab7f3c1-eee7-4a3c-985a-91ec5b148243 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 732.337287] env[63175]: DEBUG nova.compute.manager [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Refreshing instance network info cache due to event network-changed-bab7f3c1-eee7-4a3c-985a-91ec5b148243. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 732.337287] env[63175]: DEBUG oslo_concurrency.lockutils [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] Acquiring lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.406742] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.584068] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Releasing lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.584559] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 732.584789] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.585148] env[63175]: DEBUG oslo_concurrency.lockutils [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] Acquired lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.585360] env[63175]: DEBUG nova.network.neutron [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Refreshing network info cache for port bab7f3c1-eee7-4a3c-985a-91ec5b148243 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 732.586415] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a13d66c-15c4-4276-b574-dce823ab2072 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.596461] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e899539d-f489-4992-9394-28f31bd3cfda {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.608163] env[63175]: DEBUG nova.scheduler.client.report [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 732.624791] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 295e958d-20d4-493c-856b-71880cce8b98 could not be found. [ 732.625359] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.625359] env[63175]: INFO nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Took 0.04 seconds to destroy the instance on the hypervisor. [ 732.625494] env[63175]: DEBUG oslo.service.loopingcall [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.625752] env[63175]: DEBUG nova.compute.manager [-] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.625847] env[63175]: DEBUG nova.network.neutron [-] [instance: 295e958d-20d4-493c-856b-71880cce8b98] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.653360] env[63175]: DEBUG nova.network.neutron [-] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.908558] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: 223ccdbc-cee3-4040-a1ee-a53929025926] Took 1.03 seconds to deallocate network for instance. [ 733.104773] env[63175]: DEBUG nova.network.neutron [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.114429] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.115195] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.118211] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.410s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.118737] env[63175]: INFO nova.compute.claims [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.157076] env[63175]: DEBUG nova.network.neutron [-] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.203445] env[63175]: DEBUG nova.network.neutron [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.624266] env[63175]: DEBUG nova.compute.utils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.627435] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 733.627435] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.659427] env[63175]: INFO nova.compute.manager [-] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Took 1.03 seconds to deallocate network for instance. [ 733.662181] env[63175]: DEBUG nova.compute.claims [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.662181] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.682561] env[63175]: DEBUG nova.policy [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6383b3298134c2c85ce9ec51e20489d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ddc84441dcc4208b3978cdc56f86e04', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.705821] env[63175]: DEBUG oslo_concurrency.lockutils [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] Releasing lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.705912] env[63175]: DEBUG nova.compute.manager [req-1ed4bc11-1884-4a6c-90bf-4455030ba26e req-7e5f0e55-d4bc-433e-ace1-1aa40c45054b service nova] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Received event network-vif-deleted-bab7f3c1-eee7-4a3c-985a-91ec5b148243 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 733.940286] env[63175]: INFO nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Deleted allocations for instance 223ccdbc-cee3-4040-a1ee-a53929025926 [ 734.128205] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.186680] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Successfully created port: e7cccf66-6bb0-4030-8e8c-86230459af04 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.352340] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.354702] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.446982] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfba61d-8421-4f6c-a739-f4ec96df818a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.453131] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "223ccdbc-cee3-4040-a1ee-a53929025926" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.993s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.460881] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15098316-a2ce-4ab2-be4e-f767dfc1ab76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.507426] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373a0fd5-30bb-46cf-9336-989ed35901b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.514983] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4e495f-79c0-48c3-b139-6a2df5ef7c07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.528136] env[63175]: DEBUG nova.compute.provider_tree [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.861602] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.862103] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 734.862103] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 734.940507] env[63175]: DEBUG nova.compute.manager [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Received event network-changed-e7cccf66-6bb0-4030-8e8c-86230459af04 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 734.940507] env[63175]: DEBUG nova.compute.manager [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Refreshing instance network info cache due to event network-changed-e7cccf66-6bb0-4030-8e8c-86230459af04. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 734.940507] env[63175]: DEBUG oslo_concurrency.lockutils [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] Acquiring lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.940507] env[63175]: DEBUG oslo_concurrency.lockutils [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] Acquired lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.940507] env[63175]: DEBUG nova.network.neutron [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Refreshing network info cache for port e7cccf66-6bb0-4030-8e8c-86230459af04 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 734.955459] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 735.033021] env[63175]: DEBUG nova.scheduler.client.report [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 735.106272] env[63175]: ERROR nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 735.106272] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.106272] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.106272] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.106272] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.106272] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.106272] env[63175]: ERROR nova.compute.manager raise self.value [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.106272] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.106272] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.106272] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.107022] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.107022] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.107022] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 735.107022] env[63175]: ERROR nova.compute.manager [ 735.107022] env[63175]: Traceback (most recent call last): [ 735.107022] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.107022] env[63175]: listener.cb(fileno) [ 735.107022] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.107022] env[63175]: result = function(*args, **kwargs) [ 735.107022] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.107022] env[63175]: return func(*args, **kwargs) [ 735.107022] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 735.107022] env[63175]: raise e [ 735.107022] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.107022] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 735.107022] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.107022] env[63175]: created_port_ids = self._update_ports_for_instance( [ 735.107022] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.107022] env[63175]: with excutils.save_and_reraise_exception(): [ 735.107022] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.107022] env[63175]: self.force_reraise() [ 735.107022] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.107022] env[63175]: raise self.value [ 735.107022] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.107022] env[63175]: updated_port = self._update_port( [ 735.107022] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.107022] env[63175]: _ensure_no_port_binding_failure(port) [ 735.107022] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.107022] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.108273] env[63175]: nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 735.108273] env[63175]: Removing descriptor: 15 [ 735.140321] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.167828] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.168089] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.168254] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.168438] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.168585] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.168729] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.168940] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.169114] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.169285] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.169459] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.169635] env[63175]: DEBUG nova.virt.hardware [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.170496] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90969aa6-e38f-4887-aab4-150ab7fdd528 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.178289] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f955b0-2ac1-4678-904f-3fbfd4b3862c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.193991] env[63175]: ERROR nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Traceback (most recent call last): [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] yield resources [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.driver.spawn(context, instance, image_meta, [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] vm_ref = self.build_virtual_machine(instance, [ 735.193991] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] for vif in network_info: [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return self._sync_wrapper(fn, *args, **kwargs) [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.wait() [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self[:] = self._gt.wait() [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return self._exit_event.wait() [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.194347] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] current.throw(*self._exc) [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] result = function(*args, **kwargs) [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return func(*args, **kwargs) [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise e [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] nwinfo = self.network_api.allocate_for_instance( [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] created_port_ids = self._update_ports_for_instance( [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] with excutils.save_and_reraise_exception(): [ 735.194739] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.force_reraise() [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise self.value [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] updated_port = self._update_port( [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] _ensure_no_port_binding_failure(port) [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise exception.PortBindingFailed(port_id=port['id']) [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 735.195337] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] [ 735.195337] env[63175]: INFO nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Terminating instance [ 735.365122] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 735.365302] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 735.365435] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 735.365560] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 735.365680] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 735.365884] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 735.366107] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.366288] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.366502] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.366676] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.366819] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.366961] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.367098] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 735.367244] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.459270] env[63175]: DEBUG nova.network.neutron [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.479423] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.510970] env[63175]: DEBUG nova.network.neutron [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.536037] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.536037] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 735.538691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.834s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.540129] env[63175]: INFO nova.compute.claims [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.702715] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquiring lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.870499] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.013953] env[63175]: DEBUG oslo_concurrency.lockutils [req-3119dda5-57b4-47a1-9480-b6e2931654f0 req-3e7c66fc-03f7-44bf-8e8d-f69528dc2b68 service nova] Releasing lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.014838] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquired lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.014838] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.046899] env[63175]: DEBUG nova.compute.utils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.048589] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 736.048759] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.093276] env[63175]: DEBUG nova.policy [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37829cecf019424daf71f243a1e41556', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b6083764b84499a64613d988d4c264', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.412828] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Successfully created port: 9882fd21-25dd-46b9-b5eb-be23428e04e3 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.537519] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.555679] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 736.600592] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.931400] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41c988f-0614-4104-8fae-3d52b23f5880 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.941233] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989b28ea-8a3c-4e77-b100-069a65aebe04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.976454] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c8f174-07ca-44d1-8c5c-84f612a4f744 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.984666] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c05a08-26fd-444f-a203-448726eac982 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.990892] env[63175]: DEBUG nova.compute.manager [req-3bd5ec58-c020-404f-819b-506eea16ae4a req-a0a0dee7-c824-4685-b267-1b6f7969acc8 service nova] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Received event network-vif-deleted-e7cccf66-6bb0-4030-8e8c-86230459af04 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 737.002426] env[63175]: DEBUG nova.compute.provider_tree [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.063017] env[63175]: INFO nova.virt.block_device [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Booting with volume 2cac91cf-b514-4cf3-8374-df993320dbf5 at /dev/sda [ 737.106648] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Releasing lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.107282] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 737.107602] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.108055] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd5f2fc1-04d1-4243-a2e5-a2fdf0987834 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.120605] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56765a5f-b677-4780-a721-278b94b2b88c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.134219] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-158534bd-5da8-4863-ac03-329dad053cef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.142073] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc52bc5-8115-47d9-9195-78a821c044b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.156744] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4dcb8813-c428-4e0f-a761-3b06a8cba173 could not be found. [ 737.156869] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.157496] env[63175]: INFO nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Took 0.05 seconds to destroy the instance on the hypervisor. [ 737.157496] env[63175]: DEBUG oslo.service.loopingcall [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.157948] env[63175]: DEBUG nova.compute.manager [-] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 737.158084] env[63175]: DEBUG nova.network.neutron [-] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.174383] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fae7195d-5713-42f3-8662-87bf54d15736 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.185410] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada8cdec-dd6a-434d-9f9c-8f78dd133569 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.196419] env[63175]: DEBUG nova.network.neutron [-] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.212233] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f4c932-bcc7-461f-acfe-4c13f3120fc4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.219520] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e8acad-6251-432d-8e2f-acade1625001 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.233558] env[63175]: DEBUG nova.virt.block_device [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating existing volume attachment record: 88ae276e-fef6-40af-8d59-3e45dbb1fe8b {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 737.509068] env[63175]: DEBUG nova.scheduler.client.report [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 737.516030] env[63175]: ERROR nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 737.516030] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.516030] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.516030] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.516030] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.516030] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.516030] env[63175]: ERROR nova.compute.manager raise self.value [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.516030] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.516030] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.516030] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.516567] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.516567] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.516567] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 737.516567] env[63175]: ERROR nova.compute.manager [ 737.516567] env[63175]: Traceback (most recent call last): [ 737.516567] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.516567] env[63175]: listener.cb(fileno) [ 737.516567] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.516567] env[63175]: result = function(*args, **kwargs) [ 737.516567] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.516567] env[63175]: return func(*args, **kwargs) [ 737.516567] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 737.516567] env[63175]: raise e [ 737.516567] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.516567] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 737.516567] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.516567] env[63175]: created_port_ids = self._update_ports_for_instance( [ 737.516567] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.516567] env[63175]: with excutils.save_and_reraise_exception(): [ 737.516567] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.516567] env[63175]: self.force_reraise() [ 737.516567] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.516567] env[63175]: raise self.value [ 737.516567] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.516567] env[63175]: updated_port = self._update_port( [ 737.516567] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.516567] env[63175]: _ensure_no_port_binding_failure(port) [ 737.516567] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.516567] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.517500] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 737.517500] env[63175]: Removing descriptor: 15 [ 737.699078] env[63175]: DEBUG nova.network.neutron [-] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.012101] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.014056] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 738.016280] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.461s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.202157] env[63175]: INFO nova.compute.manager [-] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Took 1.04 seconds to deallocate network for instance. [ 738.204821] env[63175]: DEBUG nova.compute.claims [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.205021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.524088] env[63175]: DEBUG nova.compute.utils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.525691] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 738.525864] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 738.578008] env[63175]: DEBUG nova.policy [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f5403fbaaad420ab6c993d59ad3f294', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd5859ef0d1d4f36b0b210c2d0e84b75', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.815210] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534287dc-d0ef-4026-8fba-bc1369aa40d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.824287] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9fcc6b-6a9d-4cca-a24f-d5ce090e3a5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.860396] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edefb39-31da-42cd-a866-16a32d6168b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.866076] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cee51fb-f0e1-4ffe-be75-1d7bb990eaf7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.880517] env[63175]: DEBUG nova.compute.provider_tree [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.882162] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Successfully created port: bb8ecdac-3344-4b00-a5e4-b40d50ec974c {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.029498] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 739.034448] env[63175]: DEBUG nova.compute.manager [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Received event network-changed-9882fd21-25dd-46b9-b5eb-be23428e04e3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 739.034676] env[63175]: DEBUG nova.compute.manager [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Refreshing instance network info cache due to event network-changed-9882fd21-25dd-46b9-b5eb-be23428e04e3. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 739.034921] env[63175]: DEBUG oslo_concurrency.lockutils [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] Acquiring lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.035107] env[63175]: DEBUG oslo_concurrency.lockutils [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] Acquired lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.035343] env[63175]: DEBUG nova.network.neutron [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Refreshing network info cache for port 9882fd21-25dd-46b9-b5eb-be23428e04e3 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.359977] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 739.360550] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.360764] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.360916] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.361115] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.361321] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.361578] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.361890] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.362131] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.362340] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.362520] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.362691] env[63175]: DEBUG nova.virt.hardware [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.363598] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc9271b-0330-43cb-ac98-f9dff122b3e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.372309] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0adb3a-9281-418d-b005-205afc413f02 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.387432] env[63175]: DEBUG nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 739.391016] env[63175]: ERROR nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Traceback (most recent call last): [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] yield resources [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.driver.spawn(context, instance, image_meta, [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] vm_ref = self.build_virtual_machine(instance, [ 739.391016] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] for vif in network_info: [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return self._sync_wrapper(fn, *args, **kwargs) [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.wait() [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self[:] = self._gt.wait() [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return self._exit_event.wait() [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.391434] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] current.throw(*self._exc) [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] result = function(*args, **kwargs) [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return func(*args, **kwargs) [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise e [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] nwinfo = self.network_api.allocate_for_instance( [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] created_port_ids = self._update_ports_for_instance( [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] with excutils.save_and_reraise_exception(): [ 739.391855] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.force_reraise() [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise self.value [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] updated_port = self._update_port( [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] _ensure_no_port_binding_failure(port) [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise exception.PortBindingFailed(port_id=port['id']) [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 739.392297] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] [ 739.392297] env[63175]: INFO nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Terminating instance [ 739.565138] env[63175]: DEBUG nova.network.neutron [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.660550] env[63175]: DEBUG nova.network.neutron [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.829015] env[63175]: ERROR nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 739.829015] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.829015] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.829015] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.829015] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.829015] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.829015] env[63175]: ERROR nova.compute.manager raise self.value [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.829015] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.829015] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.829015] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.829570] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.829570] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.829570] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 739.829570] env[63175]: ERROR nova.compute.manager [ 739.829570] env[63175]: Traceback (most recent call last): [ 739.829570] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.829570] env[63175]: listener.cb(fileno) [ 739.829570] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.829570] env[63175]: result = function(*args, **kwargs) [ 739.829570] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.829570] env[63175]: return func(*args, **kwargs) [ 739.829570] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.829570] env[63175]: raise e [ 739.829570] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.829570] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 739.829570] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.829570] env[63175]: created_port_ids = self._update_ports_for_instance( [ 739.829570] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.829570] env[63175]: with excutils.save_and_reraise_exception(): [ 739.829570] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.829570] env[63175]: self.force_reraise() [ 739.829570] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.829570] env[63175]: raise self.value [ 739.829570] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.829570] env[63175]: updated_port = self._update_port( [ 739.829570] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.829570] env[63175]: _ensure_no_port_binding_failure(port) [ 739.829570] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.829570] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.830470] env[63175]: nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 739.830470] env[63175]: Removing descriptor: 15 [ 739.897688] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.898328] env[63175]: ERROR nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Traceback (most recent call last): [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.driver.spawn(context, instance, image_meta, [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] vm_ref = self.build_virtual_machine(instance, [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.898328] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] for vif in network_info: [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self._sync_wrapper(fn, *args, **kwargs) [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.wait() [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self[:] = self._gt.wait() [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self._exit_event.wait() [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] result = hub.switch() [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.898762] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return self.greenlet.switch() [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] result = function(*args, **kwargs) [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] return func(*args, **kwargs) [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise e [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] nwinfo = self.network_api.allocate_for_instance( [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] created_port_ids = self._update_ports_for_instance( [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] with excutils.save_and_reraise_exception(): [ 739.899385] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] self.force_reraise() [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise self.value [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] updated_port = self._update_port( [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] _ensure_no_port_binding_failure(port) [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] raise exception.PortBindingFailed(port_id=port['id']) [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] nova.exception.PortBindingFailed: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. [ 739.899795] env[63175]: ERROR nova.compute.manager [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] [ 739.900253] env[63175]: DEBUG nova.compute.utils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.900378] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquiring lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.900591] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.965s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.903030] env[63175]: INFO nova.compute.claims [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.906963] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Build of instance d5bc39a7-a493-41a3-a2f4-8958883adca8 was re-scheduled: Binding failed for port 305905a8-3ff7-452e-b303-c6177f0605d6, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 739.906963] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 739.906963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquiring lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.906963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Acquired lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.907355] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.045322] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 740.069495] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.069790] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.069963] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.070167] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.070313] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.070459] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.070674] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.070822] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.070982] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.071182] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.071357] env[63175]: DEBUG nova.virt.hardware [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.072717] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d421c3-3654-42b7-897e-7edde6809cf8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.080834] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d1d3d0-376e-4d1f-921f-61dca52b8b10 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.094098] env[63175]: ERROR nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Traceback (most recent call last): [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] yield resources [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.driver.spawn(context, instance, image_meta, [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] vm_ref = self.build_virtual_machine(instance, [ 740.094098] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] for vif in network_info: [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.wait() [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self[:] = self._gt.wait() [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return self._exit_event.wait() [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.094551] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] current.throw(*self._exc) [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] result = function(*args, **kwargs) [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return func(*args, **kwargs) [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise e [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] nwinfo = self.network_api.allocate_for_instance( [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] created_port_ids = self._update_ports_for_instance( [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] with excutils.save_and_reraise_exception(): [ 740.095189] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.force_reraise() [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise self.value [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] updated_port = self._update_port( [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] _ensure_no_port_binding_failure(port) [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 740.095649] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] [ 740.095649] env[63175]: INFO nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Terminating instance [ 740.163584] env[63175]: DEBUG oslo_concurrency.lockutils [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] Releasing lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.163888] env[63175]: DEBUG nova.compute.manager [req-4dd05720-0706-4dd8-b1eb-3ad548a1aaaf req-0fed4211-6fd8-46f8-b091-8b4b850da8ce service nova] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Received event network-vif-deleted-9882fd21-25dd-46b9-b5eb-be23428e04e3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 740.164264] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquired lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.164434] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.427084] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.529312] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.598694] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquiring lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.598948] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquired lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.599220] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.681656] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.759579] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.033506] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Releasing lock "refresh_cache-d5bc39a7-a493-41a3-a2f4-8958883adca8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.033766] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 741.033950] env[63175]: DEBUG nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.034129] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.050556] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.057493] env[63175]: DEBUG nova.compute.manager [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Received event network-changed-bb8ecdac-3344-4b00-a5e4-b40d50ec974c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 741.057683] env[63175]: DEBUG nova.compute.manager [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Refreshing instance network info cache due to event network-changed-bb8ecdac-3344-4b00-a5e4-b40d50ec974c. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 741.057864] env[63175]: DEBUG oslo_concurrency.lockutils [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] Acquiring lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.137849] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.217911] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310b5a02-c9ac-4702-b613-d7918665d390 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.226569] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3389b64d-ecfb-4141-9fbc-915f764a9bb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.229453] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.260707] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef6eb21-f0e9-43d3-9f98-89f36ad3d9f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.263857] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Releasing lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.264408] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 741.264689] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51d7a3ca-c3e7-42dc-b2e2-40826bbdbef3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.272017] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91dfdee-0fde-4dfd-b815-20ca9f8e0ebb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.278436] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6c8572-7776-414c-9b54-3467c1b68552 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.297669] env[63175]: DEBUG nova.compute.provider_tree [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.302398] env[63175]: WARNING nova.virt.vmwareapi.driver [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5 could not be found. [ 741.302616] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.303086] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c20174cc-37de-4dcb-a60d-a6529653b1ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.311479] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecaa1c8b-511b-44e8-869b-94a695a9a89d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.333210] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5 could not be found. [ 741.333455] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.333644] env[63175]: INFO nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Took 0.07 seconds to destroy the instance on the hypervisor. [ 741.333883] env[63175]: DEBUG oslo.service.loopingcall [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.334104] env[63175]: DEBUG nova.compute.manager [-] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.334199] env[63175]: DEBUG nova.network.neutron [-] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.351054] env[63175]: DEBUG nova.network.neutron [-] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.553852] env[63175]: DEBUG nova.network.neutron [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.732161] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Releasing lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.732602] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 741.732799] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.733111] env[63175]: DEBUG oslo_concurrency.lockutils [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] Acquired lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.733287] env[63175]: DEBUG nova.network.neutron [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Refreshing network info cache for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.734344] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ad21115-9ec7-483d-8ac2-f1a4e5f2af08 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.743375] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b189aab6-97b5-4ac5-9bbf-16520b88add8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.763977] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 501f803b-f4ea-4777-909e-7bb808628cc5 could not be found. [ 741.764170] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.764367] env[63175]: INFO nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 741.764634] env[63175]: DEBUG oslo.service.loopingcall [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.764839] env[63175]: DEBUG nova.compute.manager [-] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.764930] env[63175]: DEBUG nova.network.neutron [-] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.780555] env[63175]: DEBUG nova.network.neutron [-] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.805057] env[63175]: DEBUG nova.scheduler.client.report [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 741.853665] env[63175]: DEBUG nova.network.neutron [-] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.057073] env[63175]: INFO nova.compute.manager [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] [instance: d5bc39a7-a493-41a3-a2f4-8958883adca8] Took 1.02 seconds to deallocate network for instance. [ 742.251586] env[63175]: DEBUG nova.network.neutron [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.283327] env[63175]: DEBUG nova.network.neutron [-] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.312233] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.312726] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 742.315066] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.497s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.316441] env[63175]: INFO nova.compute.claims [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.328314] env[63175]: DEBUG nova.network.neutron [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.356496] env[63175]: INFO nova.compute.manager [-] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Took 1.02 seconds to deallocate network for instance. [ 742.786497] env[63175]: INFO nova.compute.manager [-] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Took 1.02 seconds to deallocate network for instance. [ 742.789120] env[63175]: DEBUG nova.compute.claims [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 742.789339] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.820725] env[63175]: DEBUG nova.compute.utils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.823995] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 742.824136] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.831255] env[63175]: DEBUG oslo_concurrency.lockutils [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] Releasing lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.831503] env[63175]: DEBUG nova.compute.manager [req-df62a43e-f32a-4f4c-ab09-379f531b4ee5 req-8c4ba99f-939a-4263-8545-cb64955259e2 service nova] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Received event network-vif-deleted-bb8ecdac-3344-4b00-a5e4-b40d50ec974c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 742.877149] env[63175]: DEBUG nova.policy [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2be21e0e95042c3818d502b6049c642', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0cd43969209842aab3aa0b994c1b9789', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.907351] env[63175]: INFO nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Took 0.55 seconds to detach 1 volumes for instance. [ 742.909593] env[63175]: DEBUG nova.compute.claims [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 742.909769] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.089707] env[63175]: INFO nova.scheduler.client.report [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Deleted allocations for instance d5bc39a7-a493-41a3-a2f4-8958883adca8 [ 743.149892] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Successfully created port: 9270836e-cda2-4777-8111-59c478134506 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.325146] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 743.598964] env[63175]: DEBUG oslo_concurrency.lockutils [None req-073c4c72-9b7f-445a-b3d6-f970ffbee29b tempest-ListServersNegativeTestJSON-1277676503 tempest-ListServersNegativeTestJSON-1277676503-project-member] Lock "d5bc39a7-a493-41a3-a2f4-8958883adca8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.115s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.599801] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47613523-3e10-4b68-af9e-1e678edd2fa1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.609664] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbeb651-21b8-4443-9b40-15ef13a9d29b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.641523] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4279a73-339c-486e-b542-b8ca7c949d16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.648788] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c15829-1e6a-4303-8176-c6c6ae1834a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.661803] env[63175]: DEBUG nova.compute.provider_tree [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.898352] env[63175]: DEBUG nova.compute.manager [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Received event network-changed-9270836e-cda2-4777-8111-59c478134506 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 743.899295] env[63175]: DEBUG nova.compute.manager [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Refreshing instance network info cache due to event network-changed-9270836e-cda2-4777-8111-59c478134506. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 743.899568] env[63175]: DEBUG oslo_concurrency.lockutils [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] Acquiring lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.899710] env[63175]: DEBUG oslo_concurrency.lockutils [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] Acquired lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.899868] env[63175]: DEBUG nova.network.neutron [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Refreshing network info cache for port 9270836e-cda2-4777-8111-59c478134506 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.106711] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 744.164296] env[63175]: DEBUG nova.scheduler.client.report [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.260120] env[63175]: ERROR nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 744.260120] env[63175]: ERROR nova.compute.manager Traceback (most recent call last): [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.260120] env[63175]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.260120] env[63175]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.260120] env[63175]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.260120] env[63175]: ERROR nova.compute.manager self.force_reraise() [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.260120] env[63175]: ERROR nova.compute.manager raise self.value [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.260120] env[63175]: ERROR nova.compute.manager updated_port = self._update_port( [ 744.260120] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.260120] env[63175]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 744.260609] env[63175]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.260609] env[63175]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 744.260609] env[63175]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 744.260609] env[63175]: ERROR nova.compute.manager [ 744.260609] env[63175]: Traceback (most recent call last): [ 744.260609] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 744.260609] env[63175]: listener.cb(fileno) [ 744.260609] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.260609] env[63175]: result = function(*args, **kwargs) [ 744.260609] env[63175]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.260609] env[63175]: return func(*args, **kwargs) [ 744.260609] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 744.260609] env[63175]: raise e [ 744.260609] env[63175]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.260609] env[63175]: nwinfo = self.network_api.allocate_for_instance( [ 744.260609] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.260609] env[63175]: created_port_ids = self._update_ports_for_instance( [ 744.260609] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.260609] env[63175]: with excutils.save_and_reraise_exception(): [ 744.260609] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.260609] env[63175]: self.force_reraise() [ 744.260609] env[63175]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.260609] env[63175]: raise self.value [ 744.260609] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.260609] env[63175]: updated_port = self._update_port( [ 744.260609] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.260609] env[63175]: _ensure_no_port_binding_failure(port) [ 744.260609] env[63175]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.260609] env[63175]: raise exception.PortBindingFailed(port_id=port['id']) [ 744.261522] env[63175]: nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 744.261522] env[63175]: Removing descriptor: 15 [ 744.337234] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 744.361208] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.361488] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.361651] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.361833] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.361981] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.362149] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.362378] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.362553] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.362719] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.362877] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.363093] env[63175]: DEBUG nova.virt.hardware [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.364030] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4e7c01-cd41-4846-a10e-756db152bbc9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.373449] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5dff40-5c8d-4dea-ace9-8cdb14ee7caa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.386495] env[63175]: ERROR nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Traceback (most recent call last): [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] yield resources [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.driver.spawn(context, instance, image_meta, [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] vm_ref = self.build_virtual_machine(instance, [ 744.386495] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] for vif in network_info: [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return self._sync_wrapper(fn, *args, **kwargs) [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.wait() [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self[:] = self._gt.wait() [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return self._exit_event.wait() [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.386844] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] current.throw(*self._exc) [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] result = function(*args, **kwargs) [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return func(*args, **kwargs) [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise e [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] nwinfo = self.network_api.allocate_for_instance( [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] created_port_ids = self._update_ports_for_instance( [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] with excutils.save_and_reraise_exception(): [ 744.387245] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.force_reraise() [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise self.value [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] updated_port = self._update_port( [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] _ensure_no_port_binding_failure(port) [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise exception.PortBindingFailed(port_id=port['id']) [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 744.387666] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] [ 744.387666] env[63175]: INFO nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Terminating instance [ 744.415156] env[63175]: DEBUG nova.network.neutron [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.516066] env[63175]: DEBUG nova.network.neutron [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.628305] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.669934] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.670527] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.673145] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.866s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.674666] env[63175]: INFO nova.compute.claims [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.890986] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.018881] env[63175]: DEBUG oslo_concurrency.lockutils [req-66dceabf-0fbc-40a1-9c9f-4444264df1b3 req-776a62f1-91ca-4db9-ba1f-2d83e3102ca8 service nova] Releasing lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.019417] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.019589] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.186875] env[63175]: DEBUG nova.compute.utils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.191211] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.191524] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.247882] env[63175]: DEBUG nova.policy [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2be21e0e95042c3818d502b6049c642', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0cd43969209842aab3aa0b994c1b9789', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.559163] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.577939] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Successfully created port: 307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.692446] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.708527] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.962124] env[63175]: DEBUG nova.compute.manager [req-1ec63f56-e08a-4917-bd6d-3edd27b3296e req-1431880e-0b54-4a4e-bb99-01bf25682c2d service nova] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Received event network-vif-deleted-9270836e-cda2-4777-8111-59c478134506 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 746.069496] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bee3235-ffb7-4dbe-93e9-c9a2509b9933 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.077280] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a37d250-16c6-4994-bc11-36acee6b6422 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.111220] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e76809-0608-409a-949d-e6d74eefdff9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.118406] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d7de12-82a7-46ef-a1e0-dc6dd2b6f2fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.131249] env[63175]: DEBUG nova.compute.provider_tree [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.211711] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Releasing lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.212113] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 746.212336] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.212607] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90ef51a8-663c-4a82-ac2c-1c24fc767c1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.221344] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fd68c9-912b-469d-81bf-4f793e2cfb16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.243481] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3 could not be found. [ 746.243696] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 746.243873] env[63175]: INFO nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 746.244121] env[63175]: DEBUG oslo.service.loopingcall [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.244336] env[63175]: DEBUG nova.compute.manager [-] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 746.244431] env[63175]: DEBUG nova.network.neutron [-] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.351451] env[63175]: DEBUG nova.network.neutron [-] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.634106] env[63175]: DEBUG nova.scheduler.client.report [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 746.701550] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.726225] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.726480] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.726643] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.726819] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.726962] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.727126] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.727334] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.727485] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.727651] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.727809] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.727974] env[63175]: DEBUG nova.virt.hardware [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.728976] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d606665-c6f6-4fe2-8cbd-bbe786c809f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.736853] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0f12b9-9fc6-4e28-bbc5-b15de609c185 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.853625] env[63175]: DEBUG nova.network.neutron [-] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.139608] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.140008] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 747.143249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.607s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.287086] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Successfully updated port: 307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 747.355826] env[63175]: INFO nova.compute.manager [-] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Took 1.11 seconds to deallocate network for instance. [ 747.358258] env[63175]: DEBUG nova.compute.claims [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Aborting claim: {{(pid=63175) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 747.358437] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.649703] env[63175]: DEBUG nova.compute.utils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.655031] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 747.655031] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.693481] env[63175]: DEBUG nova.policy [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44dbdeb3607544f89c170479aff37d1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c24fc185a7b24b649148093082cba4d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.790858] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.790996] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.791164] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.912238] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c753d04-857b-4afc-a00a-402f6303c358 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.920062] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd78fec9-e26a-4852-8676-aa575526df1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.955016] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5022a9c1-dcf3-4efc-a0a7-3b25ebc6636e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.962701] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d8c3ef-68c9-4055-8748-0b040c314e1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.979710] env[63175]: DEBUG nova.compute.provider_tree [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.984319] env[63175]: DEBUG nova.compute.manager [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Received event network-vif-plugged-307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 747.984524] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Acquiring lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.984885] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.984951] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.985341] env[63175]: DEBUG nova.compute.manager [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] No waiting events found dispatching network-vif-plugged-307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 747.985542] env[63175]: WARNING nova.compute.manager [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Received unexpected event network-vif-plugged-307c257c-347b-46ac-bd04-70274084334a for instance with vm_state building and task_state spawning. [ 747.985710] env[63175]: DEBUG nova.compute.manager [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Received event network-changed-307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 747.985864] env[63175]: DEBUG nova.compute.manager [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Refreshing instance network info cache due to event network-changed-307c257c-347b-46ac-bd04-70274084334a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 747.986287] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Acquiring lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.013503] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Successfully created port: 1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.155839] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 748.328121] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.480918] env[63175]: DEBUG nova.scheduler.client.report [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 748.485079] env[63175]: DEBUG nova.network.neutron [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Updating instance_info_cache with network_info: [{"id": "307c257c-347b-46ac-bd04-70274084334a", "address": "fa:16:3e:83:45:13", "network": {"id": "ce681831-7e26-47ff-8ba3-29fb0b3fbba6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1127544816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cd43969209842aab3aa0b994c1b9789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307c257c-34", "ovs_interfaceid": "307c257c-347b-46ac-bd04-70274084334a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.662512] env[63175]: INFO nova.virt.block_device [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Booting with volume ac967057-2445-49c8-8181-4d0115179902 at /dev/sda [ 748.697797] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41c26e9c-a716-4d22-b4ae-99f63b8ed715 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.706212] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c9f66c-53fe-4463-bb6b-905769db84b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.727731] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77cfa37a-2628-4696-a636-3f04f85cd27f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.735415] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bf626f-0b76-4153-82c9-5fd1ef9b5647 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.756480] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cede329d-9f08-4326-8299-f87959eb213c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.762777] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821fb604-77a8-4180-8ae3-4180258e6bcc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.776068] env[63175]: DEBUG nova.virt.block_device [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updating existing volume attachment record: 45ed59aa-3634-428b-bc0f-c0ebab0f69b0 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 748.989008] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.846s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.989328] env[63175]: ERROR nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Traceback (most recent call last): [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.driver.spawn(context, instance, image_meta, [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] vm_ref = self.build_virtual_machine(instance, [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.989328] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] for vif in network_info: [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return self._sync_wrapper(fn, *args, **kwargs) [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.wait() [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self[:] = self._gt.wait() [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return self._exit_event.wait() [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] current.throw(*self._exc) [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.989732] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] result = function(*args, **kwargs) [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] return func(*args, **kwargs) [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise e [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] nwinfo = self.network_api.allocate_for_instance( [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] created_port_ids = self._update_ports_for_instance( [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] with excutils.save_and_reraise_exception(): [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] self.force_reraise() [ 748.990156] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise self.value [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] updated_port = self._update_port( [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] _ensure_no_port_binding_failure(port) [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] raise exception.PortBindingFailed(port_id=port['id']) [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] nova.exception.PortBindingFailed: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. [ 748.990581] env[63175]: ERROR nova.compute.manager [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] [ 748.990581] env[63175]: DEBUG nova.compute.utils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.991693] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Releasing lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.992011] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Instance network_info: |[{"id": "307c257c-347b-46ac-bd04-70274084334a", "address": "fa:16:3e:83:45:13", "network": {"id": "ce681831-7e26-47ff-8ba3-29fb0b3fbba6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1127544816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cd43969209842aab3aa0b994c1b9789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307c257c-34", "ovs_interfaceid": "307c257c-347b-46ac-bd04-70274084334a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 748.992539] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Build of instance 9af5d212-c032-4446-b3cd-87fe4e66ad7c was re-scheduled: Binding failed for port dd1e9fb1-3cbf-4653-a505-c630f36c5e9e, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 748.992996] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 748.993195] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquiring lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.993343] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Acquired lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.993502] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.994463] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.332s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.997907] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Acquired lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.998109] env[63175]: DEBUG nova.network.neutron [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Refreshing network info cache for port 307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 748.999347] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:45:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e4e52d8a-b086-4333-a5a1-938680a2d2bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '307c257c-347b-46ac-bd04-70274084334a', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.007341] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Creating folder: Project (0cd43969209842aab3aa0b994c1b9789). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 749.013675] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1caefbfa-4406-41b9-9d3f-5cb0ed03f8a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.024749] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Created folder: Project (0cd43969209842aab3aa0b994c1b9789) in parent group-v268956. [ 749.024948] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Creating folder: Instances. Parent ref: group-v268973. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 749.025204] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99e14383-6055-4a0f-89d0-473a13e5b22c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.034448] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Created folder: Instances in parent group-v268973. [ 749.034693] env[63175]: DEBUG oslo.service.loopingcall [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.034856] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 749.035283] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aea14df4-0bdc-4434-8c61-778f3d1b20f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.049240] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.057645] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.057645] env[63175]: value = "task-1247903" [ 749.057645] env[63175]: _type = "Task" [ 749.057645] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.067064] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247903, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.129465] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.285397] env[63175]: DEBUG nova.network.neutron [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Updated VIF entry in instance network info cache for port 307c257c-347b-46ac-bd04-70274084334a. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 749.285754] env[63175]: DEBUG nova.network.neutron [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Updating instance_info_cache with network_info: [{"id": "307c257c-347b-46ac-bd04-70274084334a", "address": "fa:16:3e:83:45:13", "network": {"id": "ce681831-7e26-47ff-8ba3-29fb0b3fbba6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1127544816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cd43969209842aab3aa0b994c1b9789", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e4e52d8a-b086-4333-a5a1-938680a2d2bd", "external-id": "nsx-vlan-transportzone-973", "segmentation_id": 973, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307c257c-34", "ovs_interfaceid": "307c257c-347b-46ac-bd04-70274084334a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.568877] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247903, 'name': CreateVM_Task, 'duration_secs': 0.303265} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.569069] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 749.579440] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.583023] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.583023] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 749.583023] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c901c99-4ad8-4012-9ba8-5bb6cc3d3f40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.587092] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Successfully updated port: 1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.589546] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 749.589546] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529051ab-699e-2523-4384-6df2cc3bc8f1" [ 749.589546] env[63175]: _type = "Task" [ 749.589546] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.602019] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529051ab-699e-2523-4384-6df2cc3bc8f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.634337] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Releasing lock "refresh_cache-9af5d212-c032-4446-b3cd-87fe4e66ad7c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.634337] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 749.634337] env[63175]: DEBUG nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 749.634337] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.659238] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.788591] env[63175]: DEBUG oslo_concurrency.lockutils [req-7c9ca044-a48a-4675-b613-7fa5f55fb219 req-f60cd0cb-fe10-4f04-98f3-4e8798489140 service nova] Releasing lock "refresh_cache-12a2d788-c995-468b-b2a2-17dba8ad01db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.811037] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e5ce55-e9a0-40cb-96d0-5b851c5e03fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.819012] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e193d52-f687-4fa0-bc39-c054dbb0dec8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.850240] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d224f9f6-babd-4e61-98b3-14cac6438f92 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.856833] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f9db1c-b406-42fb-8dc7-d278a70a40fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.872179] env[63175]: DEBUG nova.compute.provider_tree [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.008950] env[63175]: DEBUG nova.compute.manager [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Received event network-vif-plugged-1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 750.009297] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Acquiring lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.009530] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.009699] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.009868] env[63175]: DEBUG nova.compute.manager [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] No waiting events found dispatching network-vif-plugged-1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 750.010215] env[63175]: WARNING nova.compute.manager [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Received unexpected event network-vif-plugged-1ea57e46-d47a-423b-8def-3a0c88199865 for instance with vm_state building and task_state block_device_mapping. [ 750.010420] env[63175]: DEBUG nova.compute.manager [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Received event network-changed-1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 750.010582] env[63175]: DEBUG nova.compute.manager [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Refreshing instance network info cache due to event network-changed-1ea57e46-d47a-423b-8def-3a0c88199865. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 750.010762] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Acquiring lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.010895] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Acquired lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.011084] env[63175]: DEBUG nova.network.neutron [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Refreshing network info cache for port 1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.090979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.100786] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529051ab-699e-2523-4384-6df2cc3bc8f1, 'name': SearchDatastore_Task, 'duration_secs': 0.009724} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.101165] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.101441] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.101666] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.101811] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.101988] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.102268] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6caccdc6-1648-4d3b-994e-b14f7bfc1f07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.109805] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.110024] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 750.110733] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cde39d8-a27d-454b-b937-b16547bb40ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.115405] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 750.115405] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cabe0b-c7b5-a7f6-5f30-cd1ab4d4c5f4" [ 750.115405] env[63175]: _type = "Task" [ 750.115405] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.122651] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cabe0b-c7b5-a7f6-5f30-cd1ab4d4c5f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.161461] env[63175]: DEBUG nova.network.neutron [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.377638] env[63175]: DEBUG nova.scheduler.client.report [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.543822] env[63175]: DEBUG nova.network.neutron [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.620804] env[63175]: DEBUG nova.network.neutron [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.627920] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cabe0b-c7b5-a7f6-5f30-cd1ab4d4c5f4, 'name': SearchDatastore_Task, 'duration_secs': 0.007538} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.628685] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6eb8901b-47aa-4d98-95bc-9016e3accfbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.634490] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 750.634490] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e131e5-d973-55eb-70eb-04cbde331456" [ 750.634490] env[63175]: _type = "Task" [ 750.634490] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.642697] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e131e5-d973-55eb-70eb-04cbde331456, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.663785] env[63175]: INFO nova.compute.manager [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] [instance: 9af5d212-c032-4446-b3cd-87fe4e66ad7c] Took 1.03 seconds to deallocate network for instance. [ 750.882739] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.888s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.883368] env[63175]: ERROR nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] Traceback (most recent call last): [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.driver.spawn(context, instance, image_meta, [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] vm_ref = self.build_virtual_machine(instance, [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.883368] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] for vif in network_info: [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return self._sync_wrapper(fn, *args, **kwargs) [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.wait() [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self[:] = self._gt.wait() [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return self._exit_event.wait() [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] current.throw(*self._exc) [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.883948] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] result = function(*args, **kwargs) [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] return func(*args, **kwargs) [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise e [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] nwinfo = self.network_api.allocate_for_instance( [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] created_port_ids = self._update_ports_for_instance( [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] with excutils.save_and_reraise_exception(): [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] self.force_reraise() [ 750.884416] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise self.value [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] updated_port = self._update_port( [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] _ensure_no_port_binding_failure(port) [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] raise exception.PortBindingFailed(port_id=port['id']) [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] nova.exception.PortBindingFailed: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. [ 750.884841] env[63175]: ERROR nova.compute.manager [instance: 295e958d-20d4-493c-856b-71880cce8b98] [ 750.884841] env[63175]: DEBUG nova.compute.utils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 750.885741] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.406s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.887276] env[63175]: INFO nova.compute.claims [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.889960] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Build of instance 295e958d-20d4-493c-856b-71880cce8b98 was re-scheduled: Binding failed for port bab7f3c1-eee7-4a3c-985a-91ec5b148243, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 750.890395] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 750.890710] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquiring lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.890856] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Acquired lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.891021] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.892680] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 750.893168] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.893412] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.893531] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.894116] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.894116] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.894116] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.894294] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.894337] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.894503] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.894663] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.894835] env[63175]: DEBUG nova.virt.hardware [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.895952] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694b89f6-0ffa-4e06-97c1-8328a37a87b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.905150] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80bf8e9-5d46-4e06-b5bc-3a285e012f83 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.123246] env[63175]: DEBUG oslo_concurrency.lockutils [req-6f7d1c83-c8e6-40f5-9a0f-8278491cb715 req-ce4ca1f7-fee0-4290-991c-224cc692123c service nova] Releasing lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.123654] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquired lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.123819] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.144757] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e131e5-d973-55eb-70eb-04cbde331456, 'name': SearchDatastore_Task, 'duration_secs': 0.008876} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.145007] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.145275] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 12a2d788-c995-468b-b2a2-17dba8ad01db/12a2d788-c995-468b-b2a2-17dba8ad01db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 751.145516] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d91bf398-53ed-434e-a5fa-f0d49201c6dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.152253] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 751.152253] env[63175]: value = "task-1247904" [ 751.152253] env[63175]: _type = "Task" [ 751.152253] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.159068] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247904, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.430123] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.528732] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.653249] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.663944] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247904, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426736} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.664443] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 12a2d788-c995-468b-b2a2-17dba8ad01db/12a2d788-c995-468b-b2a2-17dba8ad01db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.664663] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.664910] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a5e20af-6ef5-46e9-b2c0-692db04c68d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.670800] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 751.670800] env[63175]: value = "task-1247905" [ 751.670800] env[63175]: _type = "Task" [ 751.670800] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.681097] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.695909] env[63175]: INFO nova.scheduler.client.report [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Deleted allocations for instance 9af5d212-c032-4446-b3cd-87fe4e66ad7c [ 751.886956] env[63175]: DEBUG nova.network.neutron [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updating instance_info_cache with network_info: [{"id": "1ea57e46-d47a-423b-8def-3a0c88199865", "address": "fa:16:3e:a7:6c:78", "network": {"id": "c96de894-df02-4bcc-802a-fb114f651b63", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1671165709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c24fc185a7b24b649148093082cba4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ea57e46-d4", "ovs_interfaceid": "1ea57e46-d47a-423b-8def-3a0c88199865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.030530] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Releasing lock "refresh_cache-295e958d-20d4-493c-856b-71880cce8b98" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.030767] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 752.030950] env[63175]: DEBUG nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 752.031136] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.166099] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7c1ff5-2bce-404c-8d73-fb166ecec9af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.190465] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056723} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.190871] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.191891] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3776b386-2a33-4fb2-9815-afebb462421c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.195596] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fbd563-02ce-4b13-807a-b215f9b81c35 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.235454] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c6e0bd7c-8d18-4221-b085-4c52c8e33a35 tempest-MigrationsAdminTest-1143453682 tempest-MigrationsAdminTest-1143453682-project-member] Lock "9af5d212-c032-4446-b3cd-87fe4e66ad7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.914s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.244355] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 12a2d788-c995-468b-b2a2-17dba8ad01db/12a2d788-c995-468b-b2a2-17dba8ad01db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.245608] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.247643] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-198df1ff-6ee9-4220-b20b-67fba67fa85b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.263976] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92aaf6c7-9532-473f-8031-0b368dea919e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.274352] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d24a32-4b42-4807-ad1d-470829164a5e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.278187] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 752.278187] env[63175]: value = "task-1247906" [ 752.278187] env[63175]: _type = "Task" [ 752.278187] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.290628] env[63175]: DEBUG nova.compute.provider_tree [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.294880] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.389612] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Releasing lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.389697] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance network_info: |[{"id": "1ea57e46-d47a-423b-8def-3a0c88199865", "address": "fa:16:3e:a7:6c:78", "network": {"id": "c96de894-df02-4bcc-802a-fb114f651b63", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1671165709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c24fc185a7b24b649148093082cba4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ea57e46-d4", "ovs_interfaceid": "1ea57e46-d47a-423b-8def-3a0c88199865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 752.390098] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:6c:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20641d67-1612-4b9c-8924-7a77df9c8e6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ea57e46-d47a-423b-8def-3a0c88199865', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.397341] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Creating folder: Project (c24fc185a7b24b649148093082cba4d3). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.397604] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f995d9c-81d0-433a-938c-25cccd5d41e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.411100] env[63175]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 752.411279] env[63175]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63175) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 752.411626] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Folder already exists: Project (c24fc185a7b24b649148093082cba4d3). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 752.411822] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Creating folder: Instances. Parent ref: group-v268969. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.412068] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff231fce-6aa0-4195-9628-e04902db249e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.421035] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Created folder: Instances in parent group-v268969. [ 752.421035] env[63175]: DEBUG oslo.service.loopingcall [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.421207] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.421411] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8a4f513-4777-40d8-af0a-8b8e4e9104bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.439861] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.439861] env[63175]: value = "task-1247909" [ 752.439861] env[63175]: _type = "Task" [ 752.439861] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.447702] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247909, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.767388] env[63175]: DEBUG nova.network.neutron [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.769591] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 752.792850] env[63175]: DEBUG nova.scheduler.client.report [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 752.797011] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247906, 'name': ReconfigVM_Task, 'duration_secs': 0.285375} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.797624] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 12a2d788-c995-468b-b2a2-17dba8ad01db/12a2d788-c995-468b-b2a2-17dba8ad01db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.799105] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63960b95-81e6-4698-ba0c-d4aa05b8f9e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.806914] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 752.806914] env[63175]: value = "task-1247910" [ 752.806914] env[63175]: _type = "Task" [ 752.806914] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.815761] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247910, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.949898] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247909, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.273754] env[63175]: INFO nova.compute.manager [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] [instance: 295e958d-20d4-493c-856b-71880cce8b98] Took 1.24 seconds to deallocate network for instance. [ 753.300137] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.300137] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 753.304386] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.305181] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.434s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.305181] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.305355] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 753.305647] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.101s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.309173] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9539a8ce-1b1c-4fa7-8c71-25583861ee39 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.323482] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ae1f10-94f0-4a0b-a37c-30763cd55c34 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.327435] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247910, 'name': Rename_Task, 'duration_secs': 0.139925} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.328368] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 753.328998] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be70887f-4988-46ba-a18c-e716114bd1bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.339933] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b0ac18-33c1-44f0-8c5e-345430a9f97c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.344055] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 753.344055] env[63175]: value = "task-1247911" [ 753.344055] env[63175]: _type = "Task" [ 753.344055] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.350550] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5092558d-c293-48b1-9a49-c2ee5aedaf16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.356975] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.393795] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181522MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 753.393795] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.450984] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247909, 'name': CreateVM_Task, 'duration_secs': 0.969032} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.450984] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.451492] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'mount_device': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268972', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'name': 'volume-ac967057-2445-49c8-8181-4d0115179902', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e137904f-dc43-4ebb-90ab-e10ea5487fe5', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'serial': 'ac967057-2445-49c8-8181-4d0115179902'}, 'attachment_id': '45ed59aa-3634-428b-bc0f-c0ebab0f69b0', 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=63175) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 753.451712] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Root volume attach. Driver type: vmdk {{(pid=63175) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 753.452551] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f456ff84-272e-4e46-8f53-d188828c9b6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.459746] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f36ced-9198-4617-a222-5925743a6119 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.465608] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba9b3cb-6055-435d-a93e-4fdcdd6c7913 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.471967] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-801b2f31-fc85-4cef-baba-a88d9a325b50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.479109] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 753.479109] env[63175]: value = "task-1247912" [ 753.479109] env[63175]: _type = "Task" [ 753.479109] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.491428] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247912, 'name': RelocateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.810768] env[63175]: DEBUG nova.compute.utils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.812975] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 753.812975] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 753.853689] env[63175]: DEBUG oslo_vmware.api [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247911, 'name': PowerOnVM_Task, 'duration_secs': 0.454631} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.853972] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.854191] env[63175]: INFO nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Took 7.15 seconds to spawn the instance on the hypervisor. [ 753.854368] env[63175]: DEBUG nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 753.855115] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e74f43d-5102-445b-bab0-cf5df254dab7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.891933] env[63175]: DEBUG nova.policy [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.996138] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247912, 'name': RelocateVM_Task, 'duration_secs': 0.365857} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.996498] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 753.996745] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268972', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'name': 'volume-ac967057-2445-49c8-8181-4d0115179902', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e137904f-dc43-4ebb-90ab-e10ea5487fe5', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'serial': 'ac967057-2445-49c8-8181-4d0115179902'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 753.997554] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a06f16-0bf9-4e4a-8894-8783cb935b2e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.017016] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228eface-9014-4286-8a19-3b9c91346a06 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.038927] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] volume-ac967057-2445-49c8-8181-4d0115179902/volume-ac967057-2445-49c8-8181-4d0115179902.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.041740] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8141a38b-1ece-421b-b04b-8e877157c3dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.061726] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 754.061726] env[63175]: value = "task-1247913" [ 754.061726] env[63175]: _type = "Task" [ 754.061726] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.071097] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.202952] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310a758f-afc1-428f-971d-3c1394adb8f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.210413] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d608169d-78ea-4540-a317-1988cfb27d58 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.240947] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d866f4-42a2-44cd-b144-11268fdd8ac9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.248547] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7938ce-43ba-4480-b739-7821ce3c9d53 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.265635] env[63175]: DEBUG nova.compute.provider_tree [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.309366] env[63175]: INFO nova.scheduler.client.report [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Deleted allocations for instance 295e958d-20d4-493c-856b-71880cce8b98 [ 754.319265] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 754.376438] env[63175]: INFO nova.compute.manager [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Took 27.57 seconds to build instance. [ 754.410902] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Successfully created port: 3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.571530] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247913, 'name': ReconfigVM_Task, 'duration_secs': 0.279436} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.571839] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Reconfigured VM instance instance-00000031 to attach disk [datastore2] volume-ac967057-2445-49c8-8181-4d0115179902/volume-ac967057-2445-49c8-8181-4d0115179902.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.576470] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55385015-884d-46dc-9759-e4ef122c87c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.591779] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 754.591779] env[63175]: value = "task-1247914" [ 754.591779] env[63175]: _type = "Task" [ 754.591779] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.600481] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247914, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.768767] env[63175]: DEBUG nova.scheduler.client.report [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 754.822911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-203d89ed-2333-4227-84df-49543b91267e tempest-ServersTestJSON-121591145 tempest-ServersTestJSON-121591145-project-member] Lock "295e958d-20d4-493c-856b-71880cce8b98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.614s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.877483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c5aedbee-0764-4c89-b79d-ecb414c22156 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.561s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.105476] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247914, 'name': ReconfigVM_Task, 'duration_secs': 0.125262} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.105630] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268972', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'name': 'volume-ac967057-2445-49c8-8181-4d0115179902', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e137904f-dc43-4ebb-90ab-e10ea5487fe5', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'serial': 'ac967057-2445-49c8-8181-4d0115179902'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 755.106150] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60884e36-c27e-4ba3-850e-932a74b34143 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.113017] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 755.113017] env[63175]: value = "task-1247915" [ 755.113017] env[63175]: _type = "Task" [ 755.113017] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.122568] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247915, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.274146] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.274875] env[63175]: ERROR nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Traceback (most recent call last): [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.driver.spawn(context, instance, image_meta, [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] vm_ref = self.build_virtual_machine(instance, [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.274875] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] for vif in network_info: [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return self._sync_wrapper(fn, *args, **kwargs) [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.wait() [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self[:] = self._gt.wait() [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return self._exit_event.wait() [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] current.throw(*self._exc) [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.275231] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] result = function(*args, **kwargs) [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] return func(*args, **kwargs) [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise e [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] nwinfo = self.network_api.allocate_for_instance( [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] created_port_ids = self._update_ports_for_instance( [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] with excutils.save_and_reraise_exception(): [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] self.force_reraise() [ 755.275592] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise self.value [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] updated_port = self._update_port( [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] _ensure_no_port_binding_failure(port) [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] raise exception.PortBindingFailed(port_id=port['id']) [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] nova.exception.PortBindingFailed: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. [ 755.275963] env[63175]: ERROR nova.compute.manager [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] [ 755.277390] env[63175]: DEBUG nova.compute.utils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.278822] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.489s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.284169] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Build of instance 4dcb8813-c428-4e0f-a761-3b06a8cba173 was re-scheduled: Binding failed for port e7cccf66-6bb0-4030-8e8c-86230459af04, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 755.284603] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 755.284844] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquiring lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.284992] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Acquired lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.285172] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.329594] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 755.332690] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.363872] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.364642] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.364884] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.365157] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.365370] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.365559] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.365805] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.365997] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.366216] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.366410] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.366629] env[63175]: DEBUG nova.virt.hardware [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.367501] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0207149a-1f40-4879-a693-5694b58bae70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.376916] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7c1608-d5b9-40dc-b27c-86cae0ae1410 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.381624] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.628210] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247915, 'name': Rename_Task, 'duration_secs': 0.135405} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.629767] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 755.629767] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df6b6ce7-25f4-4e98-bedc-14f5a8b569eb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.636547] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 755.636547] env[63175]: value = "task-1247916" [ 755.636547] env[63175]: _type = "Task" [ 755.636547] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.647103] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.830484] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.854790] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.909846] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.911382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.061931] env[63175]: DEBUG nova.compute.manager [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Received event network-vif-plugged-3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 756.063134] env[63175]: DEBUG oslo_concurrency.lockutils [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.065407] env[63175]: DEBUG oslo_concurrency.lockutils [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.065652] env[63175]: DEBUG oslo_concurrency.lockutils [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.066057] env[63175]: DEBUG nova.compute.manager [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] No waiting events found dispatching network-vif-plugged-3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 756.066799] env[63175]: WARNING nova.compute.manager [req-f86e03ae-ec93-46a0-bbcc-8a63965ace15 req-5ae2b723-174e-498e-9597-d0cc911af086 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Received unexpected event network-vif-plugged-3225ec2c-7242-41c5-ba4c-bc31cc020d93 for instance with vm_state building and task_state spawning. [ 756.154463] env[63175]: DEBUG oslo_vmware.api [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1247916, 'name': PowerOnVM_Task, 'duration_secs': 0.472854} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.154794] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 756.154947] env[63175]: INFO nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Took 5.26 seconds to spawn the instance on the hypervisor. [ 756.155131] env[63175]: DEBUG nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 756.155903] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23931465-1a84-48ab-91fc-5a9da00098cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.188019] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Successfully updated port: 3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 756.191820] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acae953-0970-4b65-acfd-064156ed2263 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.203276] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b106a7-bdb6-401a-a84e-df02dd311e9e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.237571] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607d27d3-6c1d-4be1-ab6f-6d9a6b6b5a85 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.247668] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705fee92-ba1f-4066-a071-041a731b19d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.262547] env[63175]: DEBUG nova.compute.provider_tree [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.415737] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Releasing lock "refresh_cache-4dcb8813-c428-4e0f-a761-3b06a8cba173" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.416032] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 756.416243] env[63175]: DEBUG nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 756.416438] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.440061] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.687038] env[63175]: INFO nova.compute.manager [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Took 27.89 seconds to build instance. [ 756.695770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.695770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.695770] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.765773] env[63175]: DEBUG nova.scheduler.client.report [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 756.943227] env[63175]: DEBUG nova.network.neutron [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.189257] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3557049b-28ae-43b3-86b0-a2a97f07d719 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.396s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.275217] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.275846] env[63175]: ERROR nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Traceback (most recent call last): [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.driver.spawn(context, instance, image_meta, [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] vm_ref = self.build_virtual_machine(instance, [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.275846] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] for vif in network_info: [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.wait() [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self[:] = self._gt.wait() [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return self._exit_event.wait() [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] current.throw(*self._exc) [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.276493] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] result = function(*args, **kwargs) [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] return func(*args, **kwargs) [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise e [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] nwinfo = self.network_api.allocate_for_instance( [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] created_port_ids = self._update_ports_for_instance( [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] with excutils.save_and_reraise_exception(): [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] self.force_reraise() [ 757.277437] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise self.value [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] updated_port = self._update_port( [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] _ensure_no_port_binding_failure(port) [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] nova.exception.PortBindingFailed: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. [ 757.277819] env[63175]: ERROR nova.compute.manager [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] [ 757.277819] env[63175]: DEBUG nova.compute.utils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.278133] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.368s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.280755] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Build of instance 501f803b-f4ea-4777-909e-7bb808628cc5 was re-scheduled: Binding failed for port bb8ecdac-3344-4b00-a5e4-b40d50ec974c, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 757.281260] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 757.285018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquiring lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.285018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Acquired lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.285018] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.314115] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.447344] env[63175]: INFO nova.compute.manager [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] [instance: 4dcb8813-c428-4e0f-a761-3b06a8cba173] Took 1.03 seconds to deallocate network for instance. [ 757.691694] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 757.820692] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.851048] env[63175]: DEBUG nova.network.neutron [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updating instance_info_cache with network_info: [{"id": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "address": "fa:16:3e:c5:b0:01", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3225ec2c-72", "ovs_interfaceid": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.029230] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.112196] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995bdac2-346d-48d3-94cf-8a4abfb4a3cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.120794] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42852211-397d-49c8-b65b-078280bc7176 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.156359] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cb2a86-c576-46f6-83ed-d270ae1e1312 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.165433] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9d34a9-7769-441f-aafd-fee99dcef386 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.179758] env[63175]: DEBUG nova.compute.provider_tree [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.211881] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.225789] env[63175]: DEBUG nova.compute.manager [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Received event network-changed-3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 758.225789] env[63175]: DEBUG nova.compute.manager [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Refreshing instance network info cache due to event network-changed-3225ec2c-7242-41c5-ba4c-bc31cc020d93. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 758.225789] env[63175]: DEBUG oslo_concurrency.lockutils [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] Acquiring lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.354258] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.354602] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Instance network_info: |[{"id": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "address": "fa:16:3e:c5:b0:01", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3225ec2c-72", "ovs_interfaceid": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 758.355284] env[63175]: DEBUG oslo_concurrency.lockutils [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] Acquired lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.355473] env[63175]: DEBUG nova.network.neutron [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Refreshing network info cache for port 3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.356703] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:b0:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3225ec2c-7242-41c5-ba4c-bc31cc020d93', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.368617] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating folder: Project (71ca546385844c0d803034ef9e853377). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.371451] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-314c3d6b-8995-4f32-9786-74bc3c7fe0fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.385246] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created folder: Project (71ca546385844c0d803034ef9e853377) in parent group-v268956. [ 758.385246] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating folder: Instances. Parent ref: group-v268978. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.385246] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0e5448b-d65b-4e4d-b158-ec0d2ea733c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.395021] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created folder: Instances in parent group-v268978. [ 758.395021] env[63175]: DEBUG oslo.service.loopingcall [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.395021] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.395021] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70aac9c0-060e-4607-9aaa-5567c371f698 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.414959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.415537] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.418989] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.418989] env[63175]: value = "task-1247919" [ 758.418989] env[63175]: _type = "Task" [ 758.418989] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.428671] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247919, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.488836] env[63175]: INFO nova.scheduler.client.report [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Deleted allocations for instance 4dcb8813-c428-4e0f-a761-3b06a8cba173 [ 758.532683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Releasing lock "refresh_cache-501f803b-f4ea-4777-909e-7bb808628cc5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.532976] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 758.533162] env[63175]: DEBUG nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 758.533332] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.556684] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.685889] env[63175]: DEBUG nova.scheduler.client.report [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 758.932972] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247919, 'name': CreateVM_Task, 'duration_secs': 0.391665} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.933275] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.934626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.934626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.934626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 758.934795] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9d5fa6-b579-4cba-81f3-53a98d64f990 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.940548] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 758.940548] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52336cb3-92e5-9b17-ee2d-652f0c100e7b" [ 758.940548] env[63175]: _type = "Task" [ 758.940548] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.949840] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52336cb3-92e5-9b17-ee2d-652f0c100e7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.996913] env[63175]: DEBUG oslo_concurrency.lockutils [None req-bff38f9f-124d-463a-bc36-a78c8b358ab5 tempest-TenantUsagesTestJSON-765522620 tempest-TenantUsagesTestJSON-765522620-project-member] Lock "4dcb8813-c428-4e0f-a761-3b06a8cba173" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.281s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.057113] env[63175]: DEBUG nova.network.neutron [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.191353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.191858] env[63175]: ERROR nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Traceback (most recent call last): [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.driver.spawn(context, instance, image_meta, [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] vm_ref = self.build_virtual_machine(instance, [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.191858] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] for vif in network_info: [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return self._sync_wrapper(fn, *args, **kwargs) [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.wait() [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self[:] = self._gt.wait() [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return self._exit_event.wait() [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] current.throw(*self._exc) [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.192510] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] result = function(*args, **kwargs) [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] return func(*args, **kwargs) [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise e [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] nwinfo = self.network_api.allocate_for_instance( [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] created_port_ids = self._update_ports_for_instance( [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] with excutils.save_and_reraise_exception(): [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] self.force_reraise() [ 759.193022] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise self.value [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] updated_port = self._update_port( [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] _ensure_no_port_binding_failure(port) [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] raise exception.PortBindingFailed(port_id=port['id']) [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] nova.exception.PortBindingFailed: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. [ 759.193495] env[63175]: ERROR nova.compute.manager [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] [ 759.193495] env[63175]: DEBUG nova.compute.utils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.194243] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.566s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.195701] env[63175]: INFO nova.compute.claims [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.198269] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Build of instance 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5 was re-scheduled: Binding failed for port 9882fd21-25dd-46b9-b5eb-be23428e04e3, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 759.198709] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 759.198931] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquiring lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.199105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Acquired lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.199248] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.221092] env[63175]: DEBUG nova.network.neutron [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updated VIF entry in instance network info cache for port 3225ec2c-7242-41c5-ba4c-bc31cc020d93. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 759.221513] env[63175]: DEBUG nova.network.neutron [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updating instance_info_cache with network_info: [{"id": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "address": "fa:16:3e:c5:b0:01", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3225ec2c-72", "ovs_interfaceid": "3225ec2c-7242-41c5-ba4c-bc31cc020d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.451635] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52336cb3-92e5-9b17-ee2d-652f0c100e7b, 'name': SearchDatastore_Task, 'duration_secs': 0.013809} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.451635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.451855] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.452088] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.452227] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.452409] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.452716] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51ba4c80-70ee-4740-98da-de07c83a15fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.464276] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.464276] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 759.464434] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fec2e450-63e6-4a9e-b1ac-303c2947cdc4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.469479] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 759.469479] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893063-8017-dec5-d1bd-fe121d63c65c" [ 759.469479] env[63175]: _type = "Task" [ 759.469479] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.476653] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893063-8017-dec5-d1bd-fe121d63c65c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.500189] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 759.561210] env[63175]: INFO nova.compute.manager [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] [instance: 501f803b-f4ea-4777-909e-7bb808628cc5] Took 1.03 seconds to deallocate network for instance. [ 759.720843] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.724150] env[63175]: DEBUG oslo_concurrency.lockutils [req-5c4bfb45-8c97-4204-b0d1-059e1df3a14e req-87c0110c-4d53-44af-9c3e-cede5a5dbfd1 service nova] Releasing lock "refresh_cache-0f8e580e-fb14-4db8-b995-a9ffe06d8bac" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.810735] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.979886] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893063-8017-dec5-d1bd-fe121d63c65c, 'name': SearchDatastore_Task, 'duration_secs': 0.018077} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.980645] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44dd264d-e149-4ed3-a6df-14294b856d7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.986010] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 759.986010] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c32580-8b13-f58f-e009-db2215cb9de0" [ 759.986010] env[63175]: _type = "Task" [ 759.986010] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.993312] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c32580-8b13-f58f-e009-db2215cb9de0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.023332] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.314491] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Releasing lock "refresh_cache-0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.314780] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 760.314973] env[63175]: DEBUG nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 760.315156] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.324811] env[63175]: DEBUG nova.compute.manager [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Received event network-changed-1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 760.325009] env[63175]: DEBUG nova.compute.manager [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Refreshing instance network info cache due to event network-changed-1ea57e46-d47a-423b-8def-3a0c88199865. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 760.325227] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] Acquiring lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.325365] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] Acquired lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.325519] env[63175]: DEBUG nova.network.neutron [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Refreshing network info cache for port 1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.344022] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.499318] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c32580-8b13-f58f-e009-db2215cb9de0, 'name': SearchDatastore_Task, 'duration_secs': 0.013963} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.499693] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.499803] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0f8e580e-fb14-4db8-b995-a9ffe06d8bac/0f8e580e-fb14-4db8-b995-a9ffe06d8bac.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 760.501236] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc2e4877-6e1c-4da1-8af9-7ecbe8ab43da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.512796] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 760.512796] env[63175]: value = "task-1247920" [ 760.512796] env[63175]: _type = "Task" [ 760.512796] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.526353] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.532782] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332032af-43e8-4bfe-b248-c3815ed1451b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.539933] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399ede57-41b0-4b4b-9cdd-3a3220ba2b91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.580986] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b197ae-eb96-4ed0-99ed-4a49c170759b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.590339] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec96c5ff-2aaa-4065-8d29-df1ac6f5bfb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.602136] env[63175]: DEBUG nova.compute.provider_tree [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.604181] env[63175]: INFO nova.scheduler.client.report [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Deleted allocations for instance 501f803b-f4ea-4777-909e-7bb808628cc5 [ 760.846160] env[63175]: DEBUG nova.network.neutron [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.027477] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247920, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501978} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.028699] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0f8e580e-fb14-4db8-b995-a9ffe06d8bac/0f8e580e-fb14-4db8-b995-a9ffe06d8bac.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.029102] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.029459] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cbd33df-4d1a-4264-af3a-530555becb0e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.035985] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 761.035985] env[63175]: value = "task-1247921" [ 761.035985] env[63175]: _type = "Task" [ 761.035985] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.050195] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.113425] env[63175]: DEBUG nova.scheduler.client.report [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 761.118026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-48918623-cf71-4f8d-a409-a793faadab58 tempest-ServerActionsTestJSON-1720452568 tempest-ServerActionsTestJSON-1720452568-project-member] Lock "501f803b-f4ea-4777-909e-7bb808628cc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.927s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.295714] env[63175]: DEBUG nova.network.neutron [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updated VIF entry in instance network info cache for port 1ea57e46-d47a-423b-8def-3a0c88199865. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 761.296210] env[63175]: DEBUG nova.network.neutron [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updating instance_info_cache with network_info: [{"id": "1ea57e46-d47a-423b-8def-3a0c88199865", "address": "fa:16:3e:a7:6c:78", "network": {"id": "c96de894-df02-4bcc-802a-fb114f651b63", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1671165709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c24fc185a7b24b649148093082cba4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ea57e46-d4", "ovs_interfaceid": "1ea57e46-d47a-423b-8def-3a0c88199865", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.349298] env[63175]: INFO nova.compute.manager [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] [instance: 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5] Took 1.03 seconds to deallocate network for instance. [ 761.553510] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062603} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.555118] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.555118] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e1b78a-63aa-4ff0-8a59-68b97ecc58c4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.581359] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 0f8e580e-fb14-4db8-b995-a9ffe06d8bac/0f8e580e-fb14-4db8-b995-a9ffe06d8bac.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.581784] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59a3fec3-f66c-495f-9a49-ebdc9571adaf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.603153] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 761.603153] env[63175]: value = "task-1247922" [ 761.603153] env[63175]: _type = "Task" [ 761.603153] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.612509] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247922, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.621208] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.621819] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 761.624684] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 761.630039] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.269s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.803438] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd2a16da-57ed-40c5-8e85-b51a5244f6d5 req-eee648f3-d3ea-4e24-ac18-4fae308d4e7a service nova] Releasing lock "refresh_cache-e137904f-dc43-4ebb-90ab-e10ea5487fe5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.114355] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247922, 'name': ReconfigVM_Task, 'duration_secs': 0.452374} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.114663] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 0f8e580e-fb14-4db8-b995-a9ffe06d8bac/0f8e580e-fb14-4db8-b995-a9ffe06d8bac.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.116032] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-295b8720-c11e-4fac-bb2e-95020f23b743 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.120995] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 762.120995] env[63175]: value = "task-1247923" [ 762.120995] env[63175]: _type = "Task" [ 762.120995] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.131200] env[63175]: DEBUG nova.compute.utils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.137918] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247923, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.141221] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 762.141439] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 762.155105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.203066] env[63175]: DEBUG nova.policy [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36adeea9b9fc46c9bb433af250406550', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0eb8a0c65dfd4bf6b08fa2845e29a95e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.399208] env[63175]: INFO nova.scheduler.client.report [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Deleted allocations for instance 0179d80c-9d1f-49c8-bf3a-041bb6ead8d5 [ 762.490271] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171b1340-306e-480c-bc38-8c92043a07b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.498888] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66c42e7-fd2d-4ed5-9ad8-59469a7d6fb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.533357] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5349f6e0-ff7c-4ab4-b95f-06ad4b12328a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.541729] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0af29d-1105-4513-bff8-8f9c13b902b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.555375] env[63175]: DEBUG nova.compute.provider_tree [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.632963] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247923, 'name': Rename_Task, 'duration_secs': 0.138573} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.633472] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 762.635958] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.636512] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d285694e-d118-40ec-bdf7-4d865b28c9d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.645680] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 762.645680] env[63175]: value = "task-1247924" [ 762.645680] env[63175]: _type = "Task" [ 762.645680] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.658727] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247924, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.667371] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Successfully created port: ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.761528] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "a460926e-9637-40aa-bb30-e3890a441e03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.761845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.908034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e317c621-2549-4df1-8ea5-8b15bd824cd6 tempest-ServerActionsV293TestJSON-381419268 tempest-ServerActionsV293TestJSON-381419268-project-member] Lock "0179d80c-9d1f-49c8-bf3a-041bb6ead8d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.238s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.061151] env[63175]: DEBUG nova.scheduler.client.report [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 763.160052] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247924, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.413858] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 763.570340] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.571101] env[63175]: ERROR nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Traceback (most recent call last): [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.driver.spawn(context, instance, image_meta, [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] vm_ref = self.build_virtual_machine(instance, [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.571101] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] for vif in network_info: [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return self._sync_wrapper(fn, *args, **kwargs) [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.wait() [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self[:] = self._gt.wait() [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return self._exit_event.wait() [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] current.throw(*self._exc) [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.571677] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] result = function(*args, **kwargs) [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] return func(*args, **kwargs) [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise e [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] nwinfo = self.network_api.allocate_for_instance( [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] created_port_ids = self._update_ports_for_instance( [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] with excutils.save_and_reraise_exception(): [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] self.force_reraise() [ 763.572139] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise self.value [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] updated_port = self._update_port( [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] _ensure_no_port_binding_failure(port) [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] raise exception.PortBindingFailed(port_id=port['id']) [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] nova.exception.PortBindingFailed: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. [ 763.572528] env[63175]: ERROR nova.compute.manager [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] [ 763.572528] env[63175]: DEBUG nova.compute.utils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. {{(pid=63175) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.574498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.269s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.578256] env[63175]: INFO nova.compute.claims [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.582667] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Build of instance 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3 was re-scheduled: Binding failed for port 9270836e-cda2-4777-8111-59c478134506, please check neutron logs for more information. {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 763.582667] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Unplugging VIFs for instance {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 763.582832] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.582869] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquired lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.583641] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.649803] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 763.664256] env[63175]: DEBUG oslo_vmware.api [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247924, 'name': PowerOnVM_Task, 'duration_secs': 0.750661} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.664621] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.664845] env[63175]: INFO nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Took 8.34 seconds to spawn the instance on the hypervisor. [ 763.665072] env[63175]: DEBUG nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 763.665937] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dc59f2-39a8-44a8-a3b8-028e8370e7d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.691372] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.691671] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.691832] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.692028] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.692175] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.692320] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.693598] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.693833] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.694076] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.694221] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.694411] env[63175]: DEBUG nova.virt.hardware [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.696231] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1de41e3-e10b-4154-ad89-0de26c01d7e3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.705889] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460e916d-4ca5-4da0-8129-c4297bbfbf07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.739959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.740259] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.949889] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.110858] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.190342] env[63175]: INFO nova.compute.manager [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Took 28.73 seconds to build instance. [ 764.218223] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.560049] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Successfully updated port: ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 764.692034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ec0ac8f4-64a5-4c20-91f2-b12ff159066a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.828s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.695410] env[63175]: DEBUG nova.compute.manager [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received event network-vif-plugged-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 764.696974] env[63175]: DEBUG oslo_concurrency.lockutils [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] Acquiring lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.697252] env[63175]: DEBUG oslo_concurrency.lockutils [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] Lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.697457] env[63175]: DEBUG oslo_concurrency.lockutils [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] Lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.697728] env[63175]: DEBUG nova.compute.manager [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] No waiting events found dispatching network-vif-plugged-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 764.698059] env[63175]: WARNING nova.compute.manager [req-d1394ee1-98b7-476e-9a4e-9cbc71a2cb51 req-f734c7a7-64a4-4531-8f72-c9bfa28978ca service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received unexpected event network-vif-plugged-ff40a87c-cd5f-463b-86c5-278f602298db for instance with vm_state building and task_state spawning. [ 765.493694] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Releasing lock "refresh_cache-4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.494433] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63175) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 765.494433] env[63175]: DEBUG nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 765.494433] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.496052] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.496330] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquired lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.496330] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.497035] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 765.520041] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.663108] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.663353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.764279] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13162d1-3078-4a3a-9a4e-fce5714b5a56 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.771762] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d0cc97-7d70-416c-ae8c-ce2666251513 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.807878] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e479f91-81fa-4039-85a6-384cc5c359b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.816558] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd49726-f0a8-4080-a648-8629bc6042e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.834864] env[63175]: DEBUG nova.compute.provider_tree [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.023371] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.023772] env[63175]: DEBUG nova.network.neutron [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.049387] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.169656] env[63175]: DEBUG nova.compute.utils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.338222] env[63175]: DEBUG nova.scheduler.client.report [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 766.428939] env[63175]: DEBUG nova.network.neutron [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [{"id": "ff40a87c-cd5f-463b-86c5-278f602298db", "address": "fa:16:3e:26:24:b6", "network": {"id": "d6ce3558-f0aa-432c-a89b-4b153e070b66", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1503852505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb8a0c65dfd4bf6b08fa2845e29a95e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff40a87c-cd", "ovs_interfaceid": "ff40a87c-cd5f-463b-86c5-278f602298db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.528601] env[63175]: INFO nova.compute.manager [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3] Took 1.03 seconds to deallocate network for instance. [ 766.673288] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.844218] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.271s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.844709] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 766.851846] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.461s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.934590] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Releasing lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.934590] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Instance network_info: |[{"id": "ff40a87c-cd5f-463b-86c5-278f602298db", "address": "fa:16:3e:26:24:b6", "network": {"id": "d6ce3558-f0aa-432c-a89b-4b153e070b66", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1503852505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb8a0c65dfd4bf6b08fa2845e29a95e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff40a87c-cd", "ovs_interfaceid": "ff40a87c-cd5f-463b-86c5-278f602298db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 766.934789] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:24:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff40a87c-cd5f-463b-86c5-278f602298db', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.943064] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Creating folder: Project (0eb8a0c65dfd4bf6b08fa2845e29a95e). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.943354] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53662b5b-ce90-4e97-8a08-c2b0e7493478 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.954945] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Created folder: Project (0eb8a0c65dfd4bf6b08fa2845e29a95e) in parent group-v268956. [ 766.955110] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Creating folder: Instances. Parent ref: group-v268981. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.955438] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0678f300-8d6b-41be-b54c-ea629ac6bb84 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.964488] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Created folder: Instances in parent group-v268981. [ 766.964790] env[63175]: DEBUG oslo.service.loopingcall [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.965037] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.965252] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da2a86d4-52e6-4185-9a87-7d75a8766abe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.987473] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.987473] env[63175]: value = "task-1247928" [ 766.987473] env[63175]: _type = "Task" [ 766.987473] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.995166] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247928, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.354436] env[63175]: DEBUG nova.compute.utils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.356316] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 767.499342] env[63175]: DEBUG nova.compute.manager [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 767.499625] env[63175]: DEBUG nova.compute.manager [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing instance network info cache due to event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 767.499764] env[63175]: DEBUG oslo_concurrency.lockutils [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] Acquiring lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.499906] env[63175]: DEBUG oslo_concurrency.lockutils [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] Acquired lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.500080] env[63175]: DEBUG nova.network.neutron [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.513649] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247928, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.575198] env[63175]: INFO nova.scheduler.client.report [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Deleted allocations for instance 4cef9f20-00bb-44ad-ba5f-2ea69acb95e3 [ 767.759841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.760132] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.760457] env[63175]: INFO nova.compute.manager [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Attaching volume e3bcd28e-4f1e-4056-8602-953a1ddcd605 to /dev/sdb [ 767.806969] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c15c2ee-d909-449a-8f7f-0813a30621ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.826416] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e4c283-e825-43ac-9080-f1d68ed6ea27 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.842557] env[63175]: DEBUG nova.virt.block_device [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updating existing volume attachment record: 8fb71610-2a61-46bd-b578-4ff7e17e7ea3 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 767.858781] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 767.900323] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 12a2d788-c995-468b-b2a2-17dba8ad01db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.900455] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e137904f-dc43-4ebb-90ab-e10ea5487fe5 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.900596] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0f8e580e-fb14-4db8-b995-a9ffe06d8bac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.900930] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 67e8715d-b729-4013-8cca-44eaa55c662e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.900930] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 36089589-d105-49e0-8ae7-790c814b036c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.999362] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247928, 'name': CreateVM_Task, 'duration_secs': 0.646409} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.999651] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.000412] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.000627] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.000984] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 768.001612] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2beec580-61fb-4b02-902b-5d0b84058f1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.005751] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 768.005751] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529e106d-d5e7-a0bc-a163-1c614744baa0" [ 768.005751] env[63175]: _type = "Task" [ 768.005751] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.016798] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529e106d-d5e7-a0bc-a163-1c614744baa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.084562] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8d2ce3a8-e899-4106-89c9-c0bcd0317e6f tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "4cef9f20-00bb-44ad-ba5f-2ea69acb95e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.580s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.323610] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "12a2d788-c995-468b-b2a2-17dba8ad01db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.323830] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.324382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.324899] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.325367] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.329217] env[63175]: INFO nova.compute.manager [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Terminating instance [ 768.405473] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 39e55060-73a7-4dbd-96cf-bc48d8737c1c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.411158] env[63175]: DEBUG nova.network.neutron [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updated VIF entry in instance network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 768.411158] env[63175]: DEBUG nova.network.neutron [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [{"id": "ff40a87c-cd5f-463b-86c5-278f602298db", "address": "fa:16:3e:26:24:b6", "network": {"id": "d6ce3558-f0aa-432c-a89b-4b153e070b66", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1503852505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb8a0c65dfd4bf6b08fa2845e29a95e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff40a87c-cd", "ovs_interfaceid": "ff40a87c-cd5f-463b-86c5-278f602298db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.516933] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529e106d-d5e7-a0bc-a163-1c614744baa0, 'name': SearchDatastore_Task, 'duration_secs': 0.019777} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.517326] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.517564] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.517786] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.518017] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.518119] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.518395] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c01d281f-8683-46cb-8478-c9c01d28b153 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.526068] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.526339] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.527096] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29af86cd-d316-417b-b733-80831bfa0ec8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.533118] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 768.533118] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5209ecd9-fa1a-6474-ebad-9aa265440d47" [ 768.533118] env[63175]: _type = "Task" [ 768.533118] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.540986] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5209ecd9-fa1a-6474-ebad-9aa265440d47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.587953] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 768.837540] env[63175]: DEBUG nova.compute.manager [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 768.837540] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.837540] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a3655c-b40c-4280-88ee-d17b8cfb0547 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.846568] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.846925] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69e60e29-ec7d-4d3f-83d7-aaf7a32ba88d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.857457] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 768.857457] env[63175]: value = "task-1247932" [ 768.857457] env[63175]: _type = "Task" [ 768.857457] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.867852] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.872216] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 768.905990] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.908956] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.908956] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.910961] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.911214] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.911390] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.911708] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.911913] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.912103] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.912273] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.912466] env[63175]: DEBUG nova.virt.hardware [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.913266] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance c445245b-b7d3-49c6-82c5-1e8188c89b68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.915264] env[63175]: DEBUG oslo_concurrency.lockutils [req-8929c16b-1d91-4ba1-981c-154cbdc73f48 req-27a561d4-148b-4aac-aeb5-f045c68d9ab3 service nova] Releasing lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.916549] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cb073c-6124-434a-8630-726360206e50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.927935] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9464747a-5982-4491-9b07-3d3b2ff765ae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.945650] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.951416] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Creating folder: Project (57647e8061f5456db0d0e478489fab8f). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.952087] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab5eb60e-9e7f-4f08-9eec-9b13eb71113e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.962703] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Created folder: Project (57647e8061f5456db0d0e478489fab8f) in parent group-v268956. [ 768.962921] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Creating folder: Instances. Parent ref: group-v268986. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.963196] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52b00d50-309f-4bbd-8e90-a250e7211aea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.973535] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Created folder: Instances in parent group-v268986. [ 768.973794] env[63175]: DEBUG oslo.service.loopingcall [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.974086] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 768.974303] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-983bc2b9-3de6-489d-b4b6-73aacc7361c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.993830] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.993830] env[63175]: value = "task-1247935" [ 768.993830] env[63175]: _type = "Task" [ 768.993830] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.005925] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247935, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.055093] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5209ecd9-fa1a-6474-ebad-9aa265440d47, 'name': SearchDatastore_Task, 'duration_secs': 0.010806} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.055093] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8a33635-4586-42b6-9fd4-034297ec3812 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.060330] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 769.060330] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218eb97-2f65-83e3-7cfd-85df6feb36c2" [ 769.060330] env[63175]: _type = "Task" [ 769.060330] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.072669] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218eb97-2f65-83e3-7cfd-85df6feb36c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.126667] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.369076] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247932, 'name': PowerOffVM_Task, 'duration_secs': 0.287913} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.369371] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 769.369540] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 769.369792] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eab5a9c8-f53d-4826-b912-2b9c09ba7d3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.421624] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.437528] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 769.437860] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 769.438092] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Deleting the datastore file [datastore2] 12a2d788-c995-468b-b2a2-17dba8ad01db {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.438359] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22d9b3ef-c8a7-46e3-a708-17ec217b733e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.446227] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for the task: (returnval){ [ 769.446227] env[63175]: value = "task-1247937" [ 769.446227] env[63175]: _type = "Task" [ 769.446227] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.458313] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.507581] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247935, 'name': CreateVM_Task, 'duration_secs': 0.383997} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.507828] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.508954] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.508954] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.508954] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.509084] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abbfa91d-c7d1-43fb-ba5d-651b025665b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.514191] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 769.514191] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525bf22d-8ea8-6e5a-f66c-32d12d3cf473" [ 769.514191] env[63175]: _type = "Task" [ 769.514191] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.523139] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525bf22d-8ea8-6e5a-f66c-32d12d3cf473, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.572934] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218eb97-2f65-83e3-7cfd-85df6feb36c2, 'name': SearchDatastore_Task, 'duration_secs': 0.018593} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.573168] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.573462] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 67e8715d-b729-4013-8cca-44eaa55c662e/67e8715d-b729-4013-8cca-44eaa55c662e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.573749] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-332328a7-1750-4e9a-a108-73beceb06d64 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.581208] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 769.581208] env[63175]: value = "task-1247938" [ 769.581208] env[63175]: _type = "Task" [ 769.581208] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.589256] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.926084] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a2e70590-17ed-4804-b232-57526e87d22b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.956433] env[63175]: DEBUG oslo_vmware.api [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Task: {'id': task-1247937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242665} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.956702] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.956852] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.957032] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.957204] env[63175]: INFO nova.compute.manager [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Took 1.12 seconds to destroy the instance on the hypervisor. [ 769.957457] env[63175]: DEBUG oslo.service.loopingcall [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.957651] env[63175]: DEBUG nova.compute.manager [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 769.957739] env[63175]: DEBUG nova.network.neutron [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.025390] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525bf22d-8ea8-6e5a-f66c-32d12d3cf473, 'name': SearchDatastore_Task, 'duration_secs': 0.009815} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.025705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.025939] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.026199] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.026345] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.026524] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.026789] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60385611-26a7-47e0-8e6c-4f5d29993e74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.042728] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.042728] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.042728] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9988398c-dc66-43b6-9d85-6e82e5126d5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.049478] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 770.049478] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52187d0b-4b2a-8f88-f925-dea22aad20e2" [ 770.049478] env[63175]: _type = "Task" [ 770.049478] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.059717] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52187d0b-4b2a-8f88-f925-dea22aad20e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.094343] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247938, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.407413] env[63175]: DEBUG nova.compute.manager [req-1efd3340-8bf7-4650-970f-40ef170cd14b req-79f240d7-5ed3-4ced-8f09-f90506c809d8 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Received event network-vif-deleted-307c257c-347b-46ac-bd04-70274084334a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 770.407413] env[63175]: INFO nova.compute.manager [req-1efd3340-8bf7-4650-970f-40ef170cd14b req-79f240d7-5ed3-4ced-8f09-f90506c809d8 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Neutron deleted interface 307c257c-347b-46ac-bd04-70274084334a; detaching it from the instance and deleting it from the info cache [ 770.407413] env[63175]: DEBUG nova.network.neutron [req-1efd3340-8bf7-4650-970f-40ef170cd14b req-79f240d7-5ed3-4ced-8f09-f90506c809d8 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.429052] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.561061] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52187d0b-4b2a-8f88-f925-dea22aad20e2, 'name': SearchDatastore_Task, 'duration_secs': 0.05438} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.562060] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e04657c6-4371-4c90-8170-21c5394a9360 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.569449] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 770.569449] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523dc544-c8a9-2b23-e249-3b3a04c814f9" [ 770.569449] env[63175]: _type = "Task" [ 770.569449] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.587724] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523dc544-c8a9-2b23-e249-3b3a04c814f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.597962] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598589} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.597962] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 67e8715d-b729-4013-8cca-44eaa55c662e/67e8715d-b729-4013-8cca-44eaa55c662e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.597962] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.597962] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0edafda2-63b6-4f74-8dcc-53b41b5e2f4f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.604045] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 770.604045] env[63175]: value = "task-1247940" [ 770.604045] env[63175]: _type = "Task" [ 770.604045] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.612089] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.841254] env[63175]: DEBUG nova.network.neutron [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.912183] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b97c81f8-d1df-4f57-bf15-db1414f1b286 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.919811] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d015a50-da26-49df-8610-4c0519eebfc5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.931990] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.947278] env[63175]: DEBUG nova.compute.manager [req-1efd3340-8bf7-4650-970f-40ef170cd14b req-79f240d7-5ed3-4ced-8f09-f90506c809d8 service nova] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Detach interface failed, port_id=307c257c-347b-46ac-bd04-70274084334a, reason: Instance 12a2d788-c995-468b-b2a2-17dba8ad01db could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 771.079487] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523dc544-c8a9-2b23-e249-3b3a04c814f9, 'name': SearchDatastore_Task, 'duration_secs': 0.015246} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.079763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.080028] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.080292] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c5b3c75-a724-40e2-b373-ff6a38928f66 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.086923] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 771.086923] env[63175]: value = "task-1247941" [ 771.086923] env[63175]: _type = "Task" [ 771.086923] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.094769] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247941, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.113464] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.199575} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.113737] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.114651] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7209973e-44e5-417b-a4d7-0a3ac3de4433 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.137787] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 67e8715d-b729-4013-8cca-44eaa55c662e/67e8715d-b729-4013-8cca-44eaa55c662e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.137787] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2539ef5-5c32-4026-a7cd-671621236070 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.158226] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 771.158226] env[63175]: value = "task-1247942" [ 771.158226] env[63175]: _type = "Task" [ 771.158226] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.166598] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247942, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.343811] env[63175]: INFO nova.compute.manager [-] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Took 1.39 seconds to deallocate network for instance. [ 771.436362] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 72d131a0-2617-49a1-8aff-897908929bb0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.597523] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247941, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.672933] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.852114] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.940941] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 4d07a537-7df6-4659-8760-bf7e7925da25 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.098380] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247941, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.82358} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.098642] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.098851] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.099115] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bdaa9dd-640a-4b97-b53a-6d087429bce0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.104717] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 772.104717] env[63175]: value = "task-1247943" [ 772.104717] env[63175]: _type = "Task" [ 772.104717] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.112559] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.197925] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247942, 'name': ReconfigVM_Task, 'duration_secs': 0.837476} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.197925] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 67e8715d-b729-4013-8cca-44eaa55c662e/67e8715d-b729-4013-8cca-44eaa55c662e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.197925] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-633ad05a-311c-4f32-9a89-d45da84abe3e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.197925] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 772.197925] env[63175]: value = "task-1247944" [ 772.197925] env[63175]: _type = "Task" [ 772.197925] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.199391] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247944, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.407677] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 772.407913] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268985', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'name': 'volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f8e580e-fb14-4db8-b995-a9ffe06d8bac', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'serial': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 772.408806] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda4b304-ea9d-40ff-ae38-0305a1e2b4e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.424172] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a2371d-eb8f-4047-af76-4390d28c1ac4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.447361] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605/volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.448065] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.449237] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-557c0771-4177-46b3-b155-9a7c6610e2de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.468847] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 772.468847] env[63175]: value = "task-1247945" [ 772.468847] env[63175]: _type = "Task" [ 772.468847] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.479885] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247945, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.615028] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054662} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.615345] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.616129] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddee4f1-8eca-42ae-bb03-f413f404c5c6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.638333] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.638662] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-094dd2ec-5643-49b9-9891-15be47853cfe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.658529] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 772.658529] env[63175]: value = "task-1247946" [ 772.658529] env[63175]: _type = "Task" [ 772.658529] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.666432] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.685922] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247944, 'name': Rename_Task, 'duration_secs': 0.133723} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.685922] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.685922] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-134b566b-0aa5-4686-a555-f5a7db397488 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.691630] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 772.691630] env[63175]: value = "task-1247947" [ 772.691630] env[63175]: _type = "Task" [ 772.691630] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.698882] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.963815] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 780989ee-98a1-4e99-9014-45de6921d4bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.979900] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.168061] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247946, 'name': ReconfigVM_Task, 'duration_secs': 0.404326} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.168414] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.169085] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e45eb7a-1a86-48a5-b90a-db057fe9a448 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.174853] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 773.174853] env[63175]: value = "task-1247948" [ 773.174853] env[63175]: _type = "Task" [ 773.174853] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.184041] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247948, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.201703] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247947, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.468043] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance cf12aac3-edbb-48eb-b431-70187a4ecda3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.479297] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247945, 'name': ReconfigVM_Task, 'duration_secs': 0.767842} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.480208] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfigured VM instance instance-00000032 to attach disk [datastore1] volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605/volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.485009] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64f38656-29bf-4432-94f4-c2e5a696a4e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.500904] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 773.500904] env[63175]: value = "task-1247949" [ 773.500904] env[63175]: _type = "Task" [ 773.500904] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.508614] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.685879] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247948, 'name': Rename_Task, 'duration_secs': 0.182295} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.685879] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.686228] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91e22500-7bd1-4ac4-87e7-bbcb0f8b1c93 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.692075] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 773.692075] env[63175]: value = "task-1247950" [ 773.692075] env[63175]: _type = "Task" [ 773.692075] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.702455] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247950, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.705539] env[63175]: DEBUG oslo_vmware.api [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247947, 'name': PowerOnVM_Task, 'duration_secs': 0.952655} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.705791] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.706015] env[63175]: INFO nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Took 10.06 seconds to spawn the instance on the hypervisor. [ 773.706260] env[63175]: DEBUG nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 773.706946] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbca567-e1d4-47ba-9df3-ac9890761d74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.970465] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.011966] env[63175]: DEBUG oslo_vmware.api [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247949, 'name': ReconfigVM_Task, 'duration_secs': 0.213789} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.012262] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268985', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'name': 'volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f8e580e-fb14-4db8-b995-a9ffe06d8bac', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'serial': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 774.201688] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247950, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.225558] env[63175]: INFO nova.compute.manager [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Took 29.61 seconds to build instance. [ 774.474031] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 46baf816-eff2-400c-b81c-0d5f3ce8d01e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.705307] env[63175]: DEBUG oslo_vmware.api [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247950, 'name': PowerOnVM_Task, 'duration_secs': 0.702977} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.705563] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.705759] env[63175]: INFO nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Took 5.83 seconds to spawn the instance on the hypervisor. [ 774.705908] env[63175]: DEBUG nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 774.706772] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80686318-3810-4516-9750-2573e375aa50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.728315] env[63175]: DEBUG oslo_concurrency.lockutils [None req-23ae57d6-ca36-4e5a-8814-a7526a54f91e tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.716s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.976539] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a460926e-9637-40aa-bb30-e3890a441e03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.048852] env[63175]: DEBUG nova.objects.instance [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'flavor' on Instance uuid 0f8e580e-fb14-4db8-b995-a9ffe06d8bac {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 775.203631] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.224451] env[63175]: INFO nova.compute.manager [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Took 21.95 seconds to build instance. [ 775.230385] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 775.479869] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.479869] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 775.479869] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 775.553915] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff7373cd-b983-4757-94b4-d0c86e6e3614 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.794s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.554815] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.351s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.558270] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.558492] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.558669] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.560439] env[63175]: INFO nova.compute.manager [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Terminating instance [ 775.726576] env[63175]: DEBUG oslo_concurrency.lockutils [None req-71e34e59-0997-4a1a-8170-304841e261a1 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.631s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.755232] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.774612] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0454eea-1873-4ca1-8043-c72dd324ec43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.782987] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00be0ce0-d535-4d66-8551-b2a318c844bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.818954] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937edaa8-0381-4ffc-8498-ba499a2f46a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.829401] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce99cd18-d7f8-4a4e-8b0b-5c356a7ebb7e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.844543] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.910614] env[63175]: INFO nova.compute.manager [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Rebuilding instance [ 775.964229] env[63175]: DEBUG nova.compute.manager [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 775.964229] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1aea6a-31ab-4588-be3e-6ac346dd76cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.066275] env[63175]: DEBUG nova.compute.manager [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 776.066520] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 776.066793] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4362aae8-87c2-4210-9ae4-4f7b043b5874 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.074373] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 776.074373] env[63175]: value = "task-1247951" [ 776.074373] env[63175]: _type = "Task" [ 776.074373] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.084063] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.232774] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 776.348184] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 776.585506] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247951, 'name': PowerOffVM_Task, 'duration_secs': 0.254287} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.585790] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 776.585988] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 776.586205] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268985', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'name': 'volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f8e580e-fb14-4db8-b995-a9ffe06d8bac', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'serial': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 776.586964] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8557e55-5bc6-4c9c-b995-c4b57dc7f9a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.612021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d9278d-54f3-4a4b-8e1a-49e259167cf1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.613904] env[63175]: DEBUG nova.compute.manager [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 776.613993] env[63175]: DEBUG nova.compute.manager [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing instance network info cache due to event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 776.614159] env[63175]: DEBUG oslo_concurrency.lockutils [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] Acquiring lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.614300] env[63175]: DEBUG oslo_concurrency.lockutils [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] Acquired lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.614456] env[63175]: DEBUG nova.network.neutron [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.623690] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e63a64-be16-4671-b27c-cdfd13097e7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.646893] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8c770d-a692-4682-99ca-504f43f71a1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.663374] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] The volume has not been displaced from its original location: [datastore1] volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605/volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 776.668829] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfiguring VM instance instance-00000032 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 776.669391] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3afaf0f9-55b1-4a39-a678-00ec7e6a2b64 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.688894] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 776.688894] env[63175]: value = "task-1247952" [ 776.688894] env[63175]: _type = "Task" [ 776.688894] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.697143] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.752836] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.856779] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 776.856779] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.005s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.856779] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.002s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.858185] env[63175]: INFO nova.compute.claims [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.861533] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.861762] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Cleaning up deleted instances {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 776.982938] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 776.983070] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b829bbd7-e7a0-4838-a97c-2a7fe06d55c9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.989821] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 776.989821] env[63175]: value = "task-1247953" [ 776.989821] env[63175]: _type = "Task" [ 776.989821] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.998857] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.198812] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247952, 'name': ReconfigVM_Task, 'duration_secs': 0.245289} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.199108] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Reconfigured VM instance instance-00000032 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 777.203790] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bfecacb-1051-4509-af40-9ca96c8cd974 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.218138] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 777.218138] env[63175]: value = "task-1247954" [ 777.218138] env[63175]: _type = "Task" [ 777.218138] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.226218] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.369232] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] There are 2 instances to clean {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 777.369621] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 56b3ba41-0406-4b69-89af-b00e87feaa89] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 777.419588] env[63175]: DEBUG nova.network.neutron [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updated VIF entry in instance network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 777.419992] env[63175]: DEBUG nova.network.neutron [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [{"id": "ff40a87c-cd5f-463b-86c5-278f602298db", "address": "fa:16:3e:26:24:b6", "network": {"id": "d6ce3558-f0aa-432c-a89b-4b153e070b66", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1503852505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb8a0c65dfd4bf6b08fa2845e29a95e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff40a87c-cd", "ovs_interfaceid": "ff40a87c-cd5f-463b-86c5-278f602298db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.501022] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247953, 'name': PowerOffVM_Task, 'duration_secs': 0.11473} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.501022] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 777.501022] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 777.501438] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fb237a-2856-4986-b6e2-690616844a9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.508081] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 777.508322] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8804db6-b4ba-4b3a-b71c-0d382d1f379b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.532391] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 777.533037] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 777.533037] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Deleting the datastore file [datastore1] 36089589-d105-49e0-8ae7-790c814b036c {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.533241] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b38fe901-9254-4cb9-aaf1-adbfc9da4198 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.540540] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 777.540540] env[63175]: value = "task-1247956" [ 777.540540] env[63175]: _type = "Task" [ 777.540540] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.549689] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.728191] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247954, 'name': ReconfigVM_Task, 'duration_secs': 0.381798} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.729020] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268985', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'name': 'volume-e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0f8e580e-fb14-4db8-b995-a9ffe06d8bac', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605', 'serial': 'e3bcd28e-4f1e-4056-8602-953a1ddcd605'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 777.729540] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 777.730934] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac3fb1b-56aa-4a6b-9cfa-9294b9e821f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.738034] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 777.738391] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e975af3d-acb3-4b34-9569-b5f2ecc2cec2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.878385] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 24ee74d8-33d2-4efa-97de-a642c89a3461] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 777.924725] env[63175]: DEBUG oslo_concurrency.lockutils [req-a937db1d-767a-488f-8377-c1c849e7752f req-d317e8ee-d7fd-45dc-a7aa-e40a6b33237e service nova] Releasing lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.018518] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 778.018739] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 778.018921] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleting the datastore file [datastore1] 0f8e580e-fb14-4db8-b995-a9ffe06d8bac {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 778.019194] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1072f70e-5863-450a-85a9-f35db3657e7b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.027170] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 778.027170] env[63175]: value = "task-1247958" [ 778.027170] env[63175]: _type = "Task" [ 778.027170] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.038728] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.048611] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.252673} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.051118] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.051324] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 778.051514] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.148441] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd08c0a-5a1d-410f-ab1c-9be40b650185 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.156265] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d820f202-0b8a-4f43-a610-46b5a83918c4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.187036] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e041d274-db79-42d2-add4-e674687f6e43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.192625] env[63175]: DEBUG nova.compute.manager [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 778.192854] env[63175]: DEBUG nova.compute.manager [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing instance network info cache due to event network-changed-ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 778.193087] env[63175]: DEBUG oslo_concurrency.lockutils [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] Acquiring lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.193234] env[63175]: DEBUG oslo_concurrency.lockutils [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] Acquired lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.193561] env[63175]: DEBUG nova.network.neutron [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Refreshing network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.198090] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cbb861-79ab-4768-a2b0-ae064ddf6ccf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.212941] env[63175]: DEBUG nova.compute.provider_tree [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.382089] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 778.382247] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Cleaning up deleted instances with incomplete migration {{(pid=63175) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 778.536722] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.718103] env[63175]: DEBUG nova.scheduler.client.report [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 778.885772] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 778.997463] env[63175]: DEBUG nova.network.neutron [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updated VIF entry in instance network info cache for port ff40a87c-cd5f-463b-86c5-278f602298db. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 778.998077] env[63175]: DEBUG nova.network.neutron [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [{"id": "ff40a87c-cd5f-463b-86c5-278f602298db", "address": "fa:16:3e:26:24:b6", "network": {"id": "d6ce3558-f0aa-432c-a89b-4b153e070b66", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1503852505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb8a0c65dfd4bf6b08fa2845e29a95e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff40a87c-cd", "ovs_interfaceid": "ff40a87c-cd5f-463b-86c5-278f602298db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.037582] env[63175]: DEBUG oslo_vmware.api [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1247958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.529381} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.037839] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.038089] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 779.038240] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 779.038427] env[63175]: INFO nova.compute.manager [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Took 2.97 seconds to destroy the instance on the hypervisor. [ 779.038661] env[63175]: DEBUG oslo.service.loopingcall [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.038850] env[63175]: DEBUG nova.compute.manager [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 779.038941] env[63175]: DEBUG nova.network.neutron [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.088229] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.088487] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.088649] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.088841] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.088990] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.089376] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.089607] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.089775] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.089943] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.090120] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.090741] env[63175]: DEBUG nova.virt.hardware [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.091610] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae8dfc8-da7e-4f19-965b-7f4ef546c6fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.103529] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c157a5dc-f350-4fbd-9d8e-9766d1858627 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.120176] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.126550] env[63175]: DEBUG oslo.service.loopingcall [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.127237] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.127479] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53b45fb4-e520-4c0a-941a-a0834fe5eab5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.145435] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.145435] env[63175]: value = "task-1247959" [ 779.145435] env[63175]: _type = "Task" [ 779.145435] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.153424] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247959, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.221790] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.222405] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 779.226337] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.315s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.227999] env[63175]: INFO nova.compute.claims [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.323018] env[63175]: DEBUG nova.compute.manager [req-f01f4667-7b14-4ad7-8a6f-df524b28b2ec req-6d24ca0e-f6dd-454c-af79-5843a6998cb0 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Received event network-vif-deleted-3225ec2c-7242-41c5-ba4c-bc31cc020d93 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 779.323248] env[63175]: INFO nova.compute.manager [req-f01f4667-7b14-4ad7-8a6f-df524b28b2ec req-6d24ca0e-f6dd-454c-af79-5843a6998cb0 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Neutron deleted interface 3225ec2c-7242-41c5-ba4c-bc31cc020d93; detaching it from the instance and deleting it from the info cache [ 779.323567] env[63175]: DEBUG nova.network.neutron [req-f01f4667-7b14-4ad7-8a6f-df524b28b2ec req-6d24ca0e-f6dd-454c-af79-5843a6998cb0 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.501264] env[63175]: DEBUG oslo_concurrency.lockutils [req-e2a7c5cd-f56a-4f3a-a0ef-8b122752a461 req-4715ce5c-f0db-4d2c-9cfd-d60df1d94bb0 service nova] Releasing lock "refresh_cache-67e8715d-b729-4013-8cca-44eaa55c662e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.599482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "67e8715d-b729-4013-8cca-44eaa55c662e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.599757] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.599994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.600194] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.600360] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.602674] env[63175]: INFO nova.compute.manager [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Terminating instance [ 779.656780] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247959, 'name': CreateVM_Task, 'duration_secs': 0.282914} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.656957] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.657434] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.657548] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.657865] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.658135] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b191bc9-5430-40be-ab0b-a794c507c747 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.662873] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 779.662873] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52842d9d-67ac-99c5-5e00-5b7d2b9bad71" [ 779.662873] env[63175]: _type = "Task" [ 779.662873] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.670301] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52842d9d-67ac-99c5-5e00-5b7d2b9bad71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.728441] env[63175]: DEBUG nova.compute.utils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.729889] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 779.730079] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.778352] env[63175]: DEBUG nova.policy [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc420d5547b745f790e65fa1cc25a48b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '605540b1524442baa6e48276b50c5537', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.799384] env[63175]: DEBUG nova.network.neutron [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.827189] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b0c4594-91c9-4541-a223-d80216728991 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.835576] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1658d09a-b026-48e7-8cc7-22caf255a1db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.859789] env[63175]: DEBUG nova.compute.manager [req-f01f4667-7b14-4ad7-8a6f-df524b28b2ec req-6d24ca0e-f6dd-454c-af79-5843a6998cb0 service nova] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Detach interface failed, port_id=3225ec2c-7242-41c5-ba4c-bc31cc020d93, reason: Instance 0f8e580e-fb14-4db8-b995-a9ffe06d8bac could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 780.069525] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Successfully created port: f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.106891] env[63175]: DEBUG nova.compute.manager [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 780.107289] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.108203] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942636fc-b804-4059-9a55-d46943f76b80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.116287] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.116527] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-688ee636-90df-431e-8c08-2451e3fd9c90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.122040] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 780.122040] env[63175]: value = "task-1247960" [ 780.122040] env[63175]: _type = "Task" [ 780.122040] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.129509] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.173081] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52842d9d-67ac-99c5-5e00-5b7d2b9bad71, 'name': SearchDatastore_Task, 'duration_secs': 0.015358} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.173334] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.173568] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.173798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.173949] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.174144] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.174401] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccebd47f-7a17-4d0b-aa01-67c3209b01dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.181572] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.181768] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 780.182594] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bf8a7b5-35ee-45c1-9931-e93d2a067ebf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.187695] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 780.187695] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526f3c2e-7aec-f71c-1864-a7c1ab7fd082" [ 780.187695] env[63175]: _type = "Task" [ 780.187695] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.194981] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526f3c2e-7aec-f71c-1864-a7c1ab7fd082, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.233100] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 780.302399] env[63175]: INFO nova.compute.manager [-] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Took 1.26 seconds to deallocate network for instance. [ 780.481094] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fbf69f-76aa-4172-823a-ded3538ebe61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.488763] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3cb046-e57c-4546-85ff-fa1eab27d951 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.517702] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c687f7-eab2-4fe0-a8ff-f9d37fb25890 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.525385] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6044808-a637-484f-a3cc-fc3c7d6d8ee1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.539367] env[63175]: DEBUG nova.compute.provider_tree [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.632201] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247960, 'name': PowerOffVM_Task, 'duration_secs': 0.177989} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.632484] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.632678] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.632923] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f23ac10-c401-43db-8e44-428741c0939d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.696558] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.696700] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.696884] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Deleting the datastore file [datastore1] 67e8715d-b729-4013-8cca-44eaa55c662e {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.700793] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3364f2cd-7580-4942-9ce1-53ebc7acb218 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.703135] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526f3c2e-7aec-f71c-1864-a7c1ab7fd082, 'name': SearchDatastore_Task, 'duration_secs': 0.010146} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.703435] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9584e7e5-60b7-4266-9cd0-8ece0debc1cc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.707853] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for the task: (returnval){ [ 780.707853] env[63175]: value = "task-1247962" [ 780.707853] env[63175]: _type = "Task" [ 780.707853] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.709237] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 780.709237] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526e6d81-1678-c118-f6f9-fabe8cc63347" [ 780.709237] env[63175]: _type = "Task" [ 780.709237] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.719047] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.723835] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526e6d81-1678-c118-f6f9-fabe8cc63347, 'name': SearchDatastore_Task, 'duration_secs': 0.008246} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.724078] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.724327] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.724657] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fa34d39-d631-4842-8248-8bb3763ea058 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.729642] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 780.729642] env[63175]: value = "task-1247963" [ 780.729642] env[63175]: _type = "Task" [ 780.729642] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.736877] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.871030] env[63175]: INFO nova.compute.manager [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Took 0.57 seconds to detach 1 volumes for instance. [ 781.042037] env[63175]: DEBUG nova.scheduler.client.report [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 781.218580] env[63175]: DEBUG oslo_vmware.api [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Task: {'id': task-1247962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129264} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.218905] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.219043] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.219216] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.219391] env[63175]: INFO nova.compute.manager [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 781.219632] env[63175]: DEBUG oslo.service.loopingcall [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.219845] env[63175]: DEBUG nova.compute.manager [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 781.219944] env[63175]: DEBUG nova.network.neutron [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.238853] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427511} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.239116] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.239329] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.239572] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37010dd8-2c5e-4229-ab4b-c16580582ef1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.244573] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 781.248043] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 781.248043] env[63175]: value = "task-1247964" [ 781.248043] env[63175]: _type = "Task" [ 781.248043] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.256944] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.273693] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.273935] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.274150] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.274394] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.274550] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.274701] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.274913] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.275084] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.275255] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.275417] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.275586] env[63175]: DEBUG nova.virt.hardware [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.276409] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e2454b-aacd-41eb-a803-668178d5be21 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.283807] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a07541-d617-4244-8cea-e14a6d2f47de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.377693] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.466030] env[63175]: DEBUG nova.compute.manager [req-0cffaacb-24a2-4b17-b71c-52860740893a req-9084a0d7-0a6c-4f02-a146-b51f70a9d725 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Received event network-vif-deleted-ff40a87c-cd5f-463b-86c5-278f602298db {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 781.466030] env[63175]: INFO nova.compute.manager [req-0cffaacb-24a2-4b17-b71c-52860740893a req-9084a0d7-0a6c-4f02-a146-b51f70a9d725 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Neutron deleted interface ff40a87c-cd5f-463b-86c5-278f602298db; detaching it from the instance and deleting it from the info cache [ 781.466182] env[63175]: DEBUG nova.network.neutron [req-0cffaacb-24a2-4b17-b71c-52860740893a req-9084a0d7-0a6c-4f02-a146-b51f70a9d725 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.499850] env[63175]: DEBUG nova.compute.manager [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Received event network-vif-plugged-f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 781.500093] env[63175]: DEBUG oslo_concurrency.lockutils [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] Acquiring lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.500309] env[63175]: DEBUG oslo_concurrency.lockutils [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.500477] env[63175]: DEBUG oslo_concurrency.lockutils [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.500642] env[63175]: DEBUG nova.compute.manager [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] No waiting events found dispatching network-vif-plugged-f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 781.500808] env[63175]: WARNING nova.compute.manager [req-0acbdcc0-9119-4c19-a2b2-2ca8e86e84b8 req-857cb1fb-3693-4191-8e3d-299352f2f129 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Received unexpected event network-vif-plugged-f14124e5-edc9-4afb-9633-b5e7d54b3921 for instance with vm_state building and task_state spawning. [ 781.547230] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.547727] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 781.550382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.339s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.551849] env[63175]: INFO nova.compute.claims [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.590397] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Successfully updated port: f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.758348] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078072} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.758619] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.759890] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73dd7784-52c4-4760-947d-4f4e956bdff8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.778978] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.779279] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfa70583-78a5-421d-a64d-f74c52dd6e5c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.799620] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 781.799620] env[63175]: value = "task-1247965" [ 781.799620] env[63175]: _type = "Task" [ 781.799620] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.807809] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247965, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.945017] env[63175]: DEBUG nova.network.neutron [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.968784] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-455a8524-afcf-43cb-94bf-c6a8bc28bbea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.979545] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b14842-6603-4ce6-aade-a6bce6f4c6f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.002579] env[63175]: DEBUG nova.compute.manager [req-0cffaacb-24a2-4b17-b71c-52860740893a req-9084a0d7-0a6c-4f02-a146-b51f70a9d725 service nova] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Detach interface failed, port_id=ff40a87c-cd5f-463b-86c5-278f602298db, reason: Instance 67e8715d-b729-4013-8cca-44eaa55c662e could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 782.056114] env[63175]: DEBUG nova.compute.utils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.059571] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 782.059805] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.097019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.097019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquired lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.097019] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.116655] env[63175]: DEBUG nova.policy [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.310423] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247965, 'name': ReconfigVM_Task, 'duration_secs': 0.331722} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.310871] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 36089589-d105-49e0-8ae7-790c814b036c/36089589-d105-49e0-8ae7-790c814b036c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.311258] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68871837-7fef-45b2-86cf-c5c6df4ef695 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.317597] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 782.317597] env[63175]: value = "task-1247966" [ 782.317597] env[63175]: _type = "Task" [ 782.317597] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.325561] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247966, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.393935] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Successfully created port: 037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.447660] env[63175]: INFO nova.compute.manager [-] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Took 1.23 seconds to deallocate network for instance. [ 782.560019] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 782.639852] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.829171] env[63175]: DEBUG nova.network.neutron [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Updating instance_info_cache with network_info: [{"id": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "address": "fa:16:3e:e0:4f:f2", "network": {"id": "555a093b-7f2b-4816-a4c5-cc69c19fcc6e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1915575734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "605540b1524442baa6e48276b50c5537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf14124e5-ed", "ovs_interfaceid": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.834148] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247966, 'name': Rename_Task, 'duration_secs': 0.152205} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.838039] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.839403] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1a76adc-1404-4171-89e8-984ef044b746 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.846684] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 782.846684] env[63175]: value = "task-1247967" [ 782.846684] env[63175]: _type = "Task" [ 782.846684] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.857066] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247967, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.957281] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.016139] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f79b6c-908c-4470-bab2-d410b894b729 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.024094] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f914923-0329-471a-a50b-2e30ee10cef4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.053466] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88893800-52d6-4b9e-9034-7fcb1b401cf2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.061108] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79bfaa2-f9ac-440a-ae31-13dae71cd455 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.077324] env[63175]: DEBUG nova.compute.provider_tree [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.335498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Releasing lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.335787] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Instance network_info: |[{"id": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "address": "fa:16:3e:e0:4f:f2", "network": {"id": "555a093b-7f2b-4816-a4c5-cc69c19fcc6e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1915575734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "605540b1524442baa6e48276b50c5537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf14124e5-ed", "ovs_interfaceid": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 783.336236] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:4f:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f14124e5-edc9-4afb-9633-b5e7d54b3921', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.345032] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Creating folder: Project (605540b1524442baa6e48276b50c5537). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 783.345032] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c6ee8c3-cb73-4f81-803b-604965b32773 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.355386] env[63175]: DEBUG oslo_vmware.api [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247967, 'name': PowerOnVM_Task, 'duration_secs': 0.452281} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.356660] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.357025] env[63175]: DEBUG nova.compute.manager [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 783.357414] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Created folder: Project (605540b1524442baa6e48276b50c5537) in parent group-v268956. [ 783.357592] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Creating folder: Instances. Parent ref: group-v268990. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 783.358293] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1012de69-3184-4261-9527-d069ac900322 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.360468] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6cad486-8ed4-4a38-9051-754dbf69651a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.368771] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Created folder: Instances in parent group-v268990. [ 783.368862] env[63175]: DEBUG oslo.service.loopingcall [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.369301] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 783.369490] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09f9b8f4-1f97-45d6-bc47-6aed6605967c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.388478] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.388478] env[63175]: value = "task-1247970" [ 783.388478] env[63175]: _type = "Task" [ 783.388478] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.395251] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247970, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.546108] env[63175]: DEBUG nova.compute.manager [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Received event network-changed-f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 783.546324] env[63175]: DEBUG nova.compute.manager [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Refreshing instance network info cache due to event network-changed-f14124e5-edc9-4afb-9633-b5e7d54b3921. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 783.546545] env[63175]: DEBUG oslo_concurrency.lockutils [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] Acquiring lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.546689] env[63175]: DEBUG oslo_concurrency.lockutils [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] Acquired lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.547296] env[63175]: DEBUG nova.network.neutron [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Refreshing network info cache for port f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 783.579989] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 783.582816] env[63175]: DEBUG nova.scheduler.client.report [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 783.610082] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.610355] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.610516] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.610701] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.610849] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.611008] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.611234] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.611394] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.611561] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.611734] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.611931] env[63175]: DEBUG nova.virt.hardware [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.613104] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf74e865-a983-4e66-abd8-44f144e1f6c4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.621789] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dede9e-b56b-446f-84a3-00f74264fe33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.876850] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.899452] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247970, 'name': CreateVM_Task, 'duration_secs': 0.312772} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.899679] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.900351] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.900594] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.900824] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.901081] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67db8dc7-2b86-4023-9936-0e75006ace6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.905839] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 783.905839] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c167be-5339-300f-9593-330112375fb8" [ 783.905839] env[63175]: _type = "Task" [ 783.905839] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.913630] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c167be-5339-300f-9593-330112375fb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.088505] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.088505] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 784.090864] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.068s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.092206] env[63175]: INFO nova.compute.claims [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.134380] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Successfully updated port: 037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.294577] env[63175]: DEBUG nova.network.neutron [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Updated VIF entry in instance network info cache for port f14124e5-edc9-4afb-9633-b5e7d54b3921. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 784.294950] env[63175]: DEBUG nova.network.neutron [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Updating instance_info_cache with network_info: [{"id": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "address": "fa:16:3e:e0:4f:f2", "network": {"id": "555a093b-7f2b-4816-a4c5-cc69c19fcc6e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1915575734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "605540b1524442baa6e48276b50c5537", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf14124e5-ed", "ovs_interfaceid": "f14124e5-edc9-4afb-9633-b5e7d54b3921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.418268] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c167be-5339-300f-9593-330112375fb8, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.418566] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.418799] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.419066] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.419183] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.419359] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.419675] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-112b5b93-4192-4095-b0c3-203d819af7e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.427616] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.427773] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 784.428460] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec42ac77-0ba0-4f35-8569-2c8fbfd80fad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.433780] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 784.433780] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5267623f-dbea-1655-1f7d-34a5db30c394" [ 784.433780] env[63175]: _type = "Task" [ 784.433780] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.440484] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5267623f-dbea-1655-1f7d-34a5db30c394, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.597605] env[63175]: DEBUG nova.compute.utils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.602253] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 784.637691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.637691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.637844] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.797870] env[63175]: DEBUG oslo_concurrency.lockutils [req-faf0c232-2c24-4a3c-8eec-c5a89fb85187 req-aedc4666-65fb-465a-a6f0-0e0c79627309 service nova] Releasing lock "refresh_cache-39e55060-73a7-4dbd-96cf-bc48d8737c1c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.838061] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "36089589-d105-49e0-8ae7-790c814b036c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.838061] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.838271] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "36089589-d105-49e0-8ae7-790c814b036c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.838451] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.838626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.841536] env[63175]: INFO nova.compute.manager [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Terminating instance [ 784.943859] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5267623f-dbea-1655-1f7d-34a5db30c394, 'name': SearchDatastore_Task, 'duration_secs': 0.008375} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.944668] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03c10827-0539-4b20-a939-ef4ee83afb29 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.949641] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 784.949641] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ffcdc3-2e78-19ff-bb20-1762bdce4f85" [ 784.949641] env[63175]: _type = "Task" [ 784.949641] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.956899] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ffcdc3-2e78-19ff-bb20-1762bdce4f85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.103545] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 785.184615] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.345815] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "refresh_cache-36089589-d105-49e0-8ae7-790c814b036c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.345815] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquired lock "refresh_cache-36089589-d105-49e0-8ae7-790c814b036c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.345815] env[63175]: DEBUG nova.network.neutron [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.362546] env[63175]: DEBUG nova.network.neutron [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.371642] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d979ba55-e5a0-41e3-ae9d-eb51c9c5126f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.379838] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9bb8071-eea2-4c3f-82d1-f2d8afedebb8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.410436] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815595db-febc-4b25-91df-a1084e012dfb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.417150] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55836d7f-1381-46e4-bf6a-922ee36a449a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.430101] env[63175]: DEBUG nova.compute.provider_tree [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.459565] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ffcdc3-2e78-19ff-bb20-1762bdce4f85, 'name': SearchDatastore_Task, 'duration_secs': 0.008505} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.459565] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.459813] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 39e55060-73a7-4dbd-96cf-bc48d8737c1c/39e55060-73a7-4dbd-96cf-bc48d8737c1c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 785.459994] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97be8948-5b26-4a8e-9945-7a054cc4dea6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.466686] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 785.466686] env[63175]: value = "task-1247971" [ 785.466686] env[63175]: _type = "Task" [ 785.466686] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.473849] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.574761] env[63175]: DEBUG nova.compute.manager [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-vif-plugged-037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 785.574761] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.574761] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.574761] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.574761] env[63175]: DEBUG nova.compute.manager [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] No waiting events found dispatching network-vif-plugged-037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 785.574946] env[63175]: WARNING nova.compute.manager [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received unexpected event network-vif-plugged-037fa23d-fc83-4fdc-810b-b97533910bbe for instance with vm_state building and task_state spawning. [ 785.574946] env[63175]: DEBUG nova.compute.manager [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-changed-037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 785.574946] env[63175]: DEBUG nova.compute.manager [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing instance network info cache due to event network-changed-037fa23d-fc83-4fdc-810b-b97533910bbe. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 785.574946] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.864942] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.865312] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Instance network_info: |[{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 785.865616] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.865798] env[63175]: DEBUG nova.network.neutron [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing network info cache for port 037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.867146] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:02:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '037fa23d-fc83-4fdc-810b-b97533910bbe', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.876039] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating folder: Project (ab8c53b470fa4c6689aef6e5d011c3b4). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.877730] env[63175]: DEBUG nova.network.neutron [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.881790] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df3534b9-4e2b-46de-86eb-5364bf332474 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.893269] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created folder: Project (ab8c53b470fa4c6689aef6e5d011c3b4) in parent group-v268956. [ 785.893269] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating folder: Instances. Parent ref: group-v268993. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.893362] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d89d0c4-3cf1-428a-8bfe-cb9ca51c1954 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.902670] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created folder: Instances in parent group-v268993. [ 785.902913] env[63175]: DEBUG oslo.service.loopingcall [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.903156] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.905185] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1826fd06-9869-4a0b-9248-4580a1cafab0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.925330] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.925330] env[63175]: value = "task-1247974" [ 785.925330] env[63175]: _type = "Task" [ 785.925330] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.935431] env[63175]: DEBUG nova.scheduler.client.report [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 785.944668] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247974, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.960490] env[63175]: DEBUG nova.network.neutron [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.979701] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423554} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.980818] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 39e55060-73a7-4dbd-96cf-bc48d8737c1c/39e55060-73a7-4dbd-96cf-bc48d8737c1c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.980818] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.981253] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b1181a6-d8a0-40fd-ab4e-f00e61e1ca85 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.989080] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 785.989080] env[63175]: value = "task-1247975" [ 785.989080] env[63175]: _type = "Task" [ 785.989080] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.006415] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.116208] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 786.138178] env[63175]: DEBUG nova.network.neutron [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updated VIF entry in instance network info cache for port 037fa23d-fc83-4fdc-810b-b97533910bbe. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 786.138178] env[63175]: DEBUG nova.network.neutron [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.143674] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.143923] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.144097] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.144287] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.144431] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.144574] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.144785] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.144945] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.145166] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.145343] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.145560] env[63175]: DEBUG nova.virt.hardware [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.146465] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be327f5-0348-4d53-8451-9c12e622946b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.155587] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03051939-0aa5-431e-b50a-d4ff9a4c5097 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.169914] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.175480] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Creating folder: Project (30fa08f8644941deba13a99a77506ce0). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 786.175790] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c75fdda-2bd1-4d18-8f87-d71311082f85 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.185180] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Created folder: Project (30fa08f8644941deba13a99a77506ce0) in parent group-v268956. [ 786.185373] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Creating folder: Instances. Parent ref: group-v268996. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 786.185868] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d2b4095-fc45-401c-a57a-65a47177a325 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.194656] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Created folder: Instances in parent group-v268996. [ 786.194885] env[63175]: DEBUG oslo.service.loopingcall [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.195083] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 786.195280] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea6b9218-fe18-49e6-af80-979997ce7b8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.211322] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.211322] env[63175]: value = "task-1247978" [ 786.211322] env[63175]: _type = "Task" [ 786.211322] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.219309] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247978, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.436812] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247974, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.442732] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.443215] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 786.445682] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.291s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.447052] env[63175]: INFO nova.compute.claims [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.463140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Releasing lock "refresh_cache-36089589-d105-49e0-8ae7-790c814b036c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.463591] env[63175]: DEBUG nova.compute.manager [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 786.463716] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.464539] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cba8fe-1007-47ac-ab2a-e3007c47d29a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.471843] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.471989] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff29c12d-6f05-4047-a312-49baec81e56e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.477741] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 786.477741] env[63175]: value = "task-1247979" [ 786.477741] env[63175]: _type = "Task" [ 786.477741] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.487326] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.498050] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062769} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.498318] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.499105] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbd2c2b-0709-430e-8d92-1ffbf21fd576 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.521188] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 39e55060-73a7-4dbd-96cf-bc48d8737c1c/39e55060-73a7-4dbd-96cf-bc48d8737c1c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.521492] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d56647f4-fbb6-4b2c-bfe1-a23bbe200700 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.539640] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 786.539640] env[63175]: value = "task-1247980" [ 786.539640] env[63175]: _type = "Task" [ 786.539640] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.547189] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.640640] env[63175]: DEBUG oslo_concurrency.lockutils [req-5654d22b-bd7c-4726-a3f8-71b2fa8717f1 req-d94d0dcf-3c41-44fd-913e-24e35cda6db5 service nova] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.720880] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247978, 'name': CreateVM_Task, 'duration_secs': 0.498826} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.721075] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.721506] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.721673] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.722124] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.722397] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4c9f85a-fc94-4577-b8e3-d1628011dc8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.726555] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 786.726555] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a3cf69-78be-e5d8-a767-705cbcda234f" [ 786.726555] env[63175]: _type = "Task" [ 786.726555] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.735809] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a3cf69-78be-e5d8-a767-705cbcda234f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.935974] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247974, 'name': CreateVM_Task, 'duration_secs': 0.510975} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.936155] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.936845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.937044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.937341] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.937589] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f6a6e4d-7ae8-445a-97e9-7dcadb943421 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.942037] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 786.942037] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529a429a-2916-96f0-d0f1-8aa7291892de" [ 786.942037] env[63175]: _type = "Task" [ 786.942037] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.949336] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529a429a-2916-96f0-d0f1-8aa7291892de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.952614] env[63175]: DEBUG nova.compute.utils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.955730] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 786.955900] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.987952] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247979, 'name': PowerOffVM_Task, 'duration_secs': 0.112446} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.988244] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.988408] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.988653] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39c032c0-143f-4009-8125-7109e8f2e472 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.995866] env[63175]: DEBUG nova.policy [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2b97dde949d4f12b46fe08c4ea19986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1847b48d1b484fdcbfc114fcf0c4b5e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.011380] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.011637] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.011796] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Deleting the datastore file [datastore2] 36089589-d105-49e0-8ae7-790c814b036c {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.012081] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-add4078d-80ab-4cab-8fde-5fc359b6f964 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.017727] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for the task: (returnval){ [ 787.017727] env[63175]: value = "task-1247982" [ 787.017727] env[63175]: _type = "Task" [ 787.017727] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.025684] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247982, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.048714] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247980, 'name': ReconfigVM_Task, 'duration_secs': 0.26551} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.049119] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 39e55060-73a7-4dbd-96cf-bc48d8737c1c/39e55060-73a7-4dbd-96cf-bc48d8737c1c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.049796] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e10bcfcb-0206-46dc-8faa-9f42e46e3449 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.055697] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 787.055697] env[63175]: value = "task-1247983" [ 787.055697] env[63175]: _type = "Task" [ 787.055697] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.063674] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247983, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.237488] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a3cf69-78be-e5d8-a767-705cbcda234f, 'name': SearchDatastore_Task, 'duration_secs': 0.007975} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.237798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.238053] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.238291] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.238439] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.238616] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.238871] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c62e53f8-d564-4a88-bad9-084f18191874 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.247209] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.247390] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.248077] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c8f162b-41ca-4069-b7ec-9040c6f776ac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.253043] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 787.253043] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e57d4f-f559-174b-8230-3da4b403ea9c" [ 787.253043] env[63175]: _type = "Task" [ 787.253043] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.260719] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e57d4f-f559-174b-8230-3da4b403ea9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.290040] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Successfully created port: 5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.452070] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529a429a-2916-96f0-d0f1-8aa7291892de, 'name': SearchDatastore_Task, 'duration_secs': 0.00989} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.452396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.452627] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.452885] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.453142] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.453262] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.453524] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9d025a5-2fb3-465e-a746-89b762991e98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.455931] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 787.473022] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.473022] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.473022] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8702555e-a7aa-4f62-9a45-b6fe50b1d3ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.475924] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 787.475924] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52069677-6238-85a7-6d2f-0225c556e670" [ 787.475924] env[63175]: _type = "Task" [ 787.475924] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.483640] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52069677-6238-85a7-6d2f-0225c556e670, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.529376] env[63175]: DEBUG oslo_vmware.api [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Task: {'id': task-1247982, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091987} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.529578] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.529754] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.530093] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.530316] env[63175]: INFO nova.compute.manager [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Took 1.07 seconds to destroy the instance on the hypervisor. [ 787.530556] env[63175]: DEBUG oslo.service.loopingcall [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.530748] env[63175]: DEBUG nova.compute.manager [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 787.530844] env[63175]: DEBUG nova.network.neutron [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.550471] env[63175]: DEBUG nova.network.neutron [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.572984] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247983, 'name': Rename_Task, 'duration_secs': 0.132046} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.573267] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.573529] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e75184d0-b01d-44fd-8236-7afa19ae5a7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.581325] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 787.581325] env[63175]: value = "task-1247984" [ 787.581325] env[63175]: _type = "Task" [ 787.581325] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.589323] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.728762] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90dc0ee6-becf-4911-8c33-403496a323b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.735962] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3473769-222b-4d2c-95bb-5d4e5ca6a1f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.767905] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104b6fa3-b907-43ea-b9ba-69aa5442eb5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.775290] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e57d4f-f559-174b-8230-3da4b403ea9c, 'name': SearchDatastore_Task, 'duration_secs': 0.007862} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.777780] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53b8f1ee-526c-4660-92fe-4bb410b4803a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.780859] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d262a84-b69c-4039-9850-644149ba7d05 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.787137] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 787.787137] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52608d25-0058-8369-9927-f229cd8bb119" [ 787.787137] env[63175]: _type = "Task" [ 787.787137] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.796043] env[63175]: DEBUG nova.compute.provider_tree [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.804245] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52608d25-0058-8369-9927-f229cd8bb119, 'name': SearchDatastore_Task, 'duration_secs': 0.0099} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.805232] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.805534] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4/35f775ef-8a9e-4c49-99b9-a90f8c8f39b4.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.805837] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bccfa18a-8443-4e45-acb3-a9c5492fe1a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.812948] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 787.812948] env[63175]: value = "task-1247985" [ 787.812948] env[63175]: _type = "Task" [ 787.812948] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.822666] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247985, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.987760] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52069677-6238-85a7-6d2f-0225c556e670, 'name': SearchDatastore_Task, 'duration_secs': 0.008501} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.988596] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce48483-fd58-40dc-b0d1-50f606129679 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.994245] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 787.994245] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e1bf05-9f6f-afcd-0b1b-e3aa835b9420" [ 787.994245] env[63175]: _type = "Task" [ 787.994245] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.002689] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e1bf05-9f6f-afcd-0b1b-e3aa835b9420, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.055189] env[63175]: DEBUG nova.network.neutron [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.092410] env[63175]: DEBUG oslo_vmware.api [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247984, 'name': PowerOnVM_Task, 'duration_secs': 0.433091} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.092410] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.092563] env[63175]: INFO nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Took 6.85 seconds to spawn the instance on the hypervisor. [ 788.092742] env[63175]: DEBUG nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 788.093593] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435c0e46-e64d-473b-a449-634b0e2f5f1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.300452] env[63175]: DEBUG nova.scheduler.client.report [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 788.324365] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247985, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443937} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.324789] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4/35f775ef-8a9e-4c49-99b9-a90f8c8f39b4.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.325145] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.325510] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4dbef7f3-9df8-4d07-a209-0288765cc66d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.333778] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 788.333778] env[63175]: value = "task-1247986" [ 788.333778] env[63175]: _type = "Task" [ 788.333778] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.343999] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.473305] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 788.501277] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.501553] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.501714] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.502017] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.502189] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.502339] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.502576] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.502743] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.502910] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.503094] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.503267] env[63175]: DEBUG nova.virt.hardware [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.504157] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228d3a37-6dc2-43f3-ad0f-7f458e6df97a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.510507] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e1bf05-9f6f-afcd-0b1b-e3aa835b9420, 'name': SearchDatastore_Task, 'duration_secs': 0.018476} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.511207] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.511478] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] c445245b-b7d3-49c6-82c5-1e8188c89b68/c445245b-b7d3-49c6-82c5-1e8188c89b68.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 788.511757] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29af14a8-2e18-45d6-bf9d-ab7ee16e2aab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.517386] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892eba92-20fc-404a-9cc3-5f90053afb6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.522259] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 788.522259] env[63175]: value = "task-1247987" [ 788.522259] env[63175]: _type = "Task" [ 788.522259] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.541193] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.558069] env[63175]: INFO nova.compute.manager [-] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Took 1.03 seconds to deallocate network for instance. [ 788.610727] env[63175]: INFO nova.compute.manager [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Took 32.77 seconds to build instance. [ 788.740138] env[63175]: DEBUG nova.compute.manager [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Received event network-vif-plugged-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 788.740400] env[63175]: DEBUG oslo_concurrency.lockutils [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] Acquiring lock "a2e70590-17ed-4804-b232-57526e87d22b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.740699] env[63175]: DEBUG oslo_concurrency.lockutils [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] Lock "a2e70590-17ed-4804-b232-57526e87d22b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.740771] env[63175]: DEBUG oslo_concurrency.lockutils [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] Lock "a2e70590-17ed-4804-b232-57526e87d22b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.741016] env[63175]: DEBUG nova.compute.manager [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] No waiting events found dispatching network-vif-plugged-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 788.741445] env[63175]: WARNING nova.compute.manager [req-af84eabd-ac97-4752-93aa-a94a126c0d3c req-f5a92b36-8358-47cb-92d1-56875e17bed8 service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Received unexpected event network-vif-plugged-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a for instance with vm_state building and task_state spawning. [ 788.806173] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.807114] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 788.811206] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.862s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.815300] env[63175]: INFO nova.compute.claims [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.846138] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149521} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.846491] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.847431] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3806a6c-e00d-47c3-940d-1458cac6c01a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.872662] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4/35f775ef-8a9e-4c49-99b9-a90f8c8f39b4.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.873430] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73030a9b-6980-4978-87c3-abd363dc8d67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.896290] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Successfully updated port: 5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.903421] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 788.903421] env[63175]: value = "task-1247988" [ 788.903421] env[63175]: _type = "Task" [ 788.903421] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.916066] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247988, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.032419] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460491} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.032694] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] c445245b-b7d3-49c6-82c5-1e8188c89b68/c445245b-b7d3-49c6-82c5-1e8188c89b68.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.032917] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.033523] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b451e21-3be7-4ece-ae70-da5fe6d50362 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.042368] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 789.042368] env[63175]: value = "task-1247989" [ 789.042368] env[63175]: _type = "Task" [ 789.042368] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.050861] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.065902] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.113039] env[63175]: DEBUG oslo_concurrency.lockutils [None req-afd6bc8b-4969-4756-a110-b15aa156629a tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.165s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.325340] env[63175]: DEBUG nova.compute.utils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.328317] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 789.328317] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.373218] env[63175]: DEBUG nova.policy [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c654f31fd4bc49dc8e985bd75a811ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74dd39e20aad4ddb874282973cea02bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.394447] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.394692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.394896] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.395091] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.395268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.397286] env[63175]: INFO nova.compute.manager [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Terminating instance [ 789.399651] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.399865] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquired lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.400125] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.413874] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247988, 'name': ReconfigVM_Task, 'duration_secs': 0.386978} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.414160] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4/35f775ef-8a9e-4c49-99b9-a90f8c8f39b4.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.414745] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ac8506e-fcb1-4ba1-9359-e15fab39488d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.421361] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 789.421361] env[63175]: value = "task-1247990" [ 789.421361] env[63175]: _type = "Task" [ 789.421361] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.430611] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247990, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.553979] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170383} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.557393] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.558419] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50ef46b-c091-4647-b8a2-ec1fbf8e72e0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.580464] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] c445245b-b7d3-49c6-82c5-1e8188c89b68/c445245b-b7d3-49c6-82c5-1e8188c89b68.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.580464] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85742501-6936-4270-bb6c-ecb49fd18729 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.598778] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 789.598778] env[63175]: value = "task-1247991" [ 789.598778] env[63175]: _type = "Task" [ 789.598778] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.606511] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.616075] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.656416] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Successfully created port: 595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.831747] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 789.901618] env[63175]: DEBUG nova.compute.manager [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 789.901873] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.904341] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e241d-092d-4b8b-b17b-2f3fcb9efe91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.915118] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 789.915118] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d3302f6-3e55-4358-84f6-39c77a0edf6c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.923150] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 789.923150] env[63175]: value = "task-1247992" [ 789.923150] env[63175]: _type = "Task" [ 789.923150] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.940196] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.940628] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247990, 'name': Rename_Task, 'duration_secs': 0.392848} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.941246] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.941677] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aef088b0-0138-48d3-8f05-e59719fdb194 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.947404] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 789.947404] env[63175]: value = "task-1247993" [ 789.947404] env[63175]: _type = "Task" [ 789.947404] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.954280] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.958899] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.112669] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.117167] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c557b7-0d52-41c2-90af-4aa1e2fbeaea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.127685] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1b3c4a-9b75-4a9d-91d6-66823038bba3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.168735] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.170108] env[63175]: DEBUG nova.network.neutron [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Updating instance_info_cache with network_info: [{"id": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "address": "fa:16:3e:60:57:ff", "network": {"id": "174eebb3-caef-4b69-89e0-15cf23f34574", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2090229537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1847b48d1b484fdcbfc114fcf0c4b5e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5301b2d9-86", "ovs_interfaceid": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.171908] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba3ee89-cab8-4727-a59c-c3e11a4bb209 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.183757] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c65be4-8a52-4b16-9e1b-e8cfc81b8e2b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.200834] env[63175]: DEBUG nova.compute.provider_tree [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.437687] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247992, 'name': PowerOffVM_Task, 'duration_secs': 0.280162} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.438776] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.438776] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.438912] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a57c8620-b65b-4447-aef4-8255ec5d4ea0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.457403] env[63175]: DEBUG oslo_vmware.api [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1247993, 'name': PowerOnVM_Task, 'duration_secs': 0.457675} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.457664] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.458357] env[63175]: INFO nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Took 4.34 seconds to spawn the instance on the hypervisor. [ 790.458357] env[63175]: DEBUG nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 790.458803] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34052af9-e2e3-41d5-b148-62d589e67679 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.503561] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.503750] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.503961] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Deleting the datastore file [datastore1] 39e55060-73a7-4dbd-96cf-bc48d8737c1c {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.504243] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dd55c07-fd52-4492-a844-6c52bef40184 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.510135] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for the task: (returnval){ [ 790.510135] env[63175]: value = "task-1247995" [ 790.510135] env[63175]: _type = "Task" [ 790.510135] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.517521] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.609989] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247991, 'name': ReconfigVM_Task, 'duration_secs': 0.649352} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.610295] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfigured VM instance instance-00000036 to attach disk [datastore1] c445245b-b7d3-49c6-82c5-1e8188c89b68/c445245b-b7d3-49c6-82c5-1e8188c89b68.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.610903] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f265c93-fd35-4e91-8548-5e8fd7dd7f45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.616745] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 790.616745] env[63175]: value = "task-1247996" [ 790.616745] env[63175]: _type = "Task" [ 790.616745] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.624852] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247996, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.676069] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Releasing lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.676419] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Instance network_info: |[{"id": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "address": "fa:16:3e:60:57:ff", "network": {"id": "174eebb3-caef-4b69-89e0-15cf23f34574", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2090229537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1847b48d1b484fdcbfc114fcf0c4b5e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5301b2d9-86", "ovs_interfaceid": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 790.676846] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:57:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '822050c7-1845-485d-b87e-73778d21c33c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5301b2d9-86e8-4b5b-abd3-6443a9d5be1a', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.684198] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Creating folder: Project (1847b48d1b484fdcbfc114fcf0c4b5e9). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.684474] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25db9216-fc96-4864-aaae-ea2660183389 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.693609] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Created folder: Project (1847b48d1b484fdcbfc114fcf0c4b5e9) in parent group-v268956. [ 790.693786] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Creating folder: Instances. Parent ref: group-v268999. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.694008] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-942ae01e-18a8-4b6a-840e-a02ea5841ead {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.703065] env[63175]: DEBUG nova.scheduler.client.report [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 790.706178] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Created folder: Instances in parent group-v268999. [ 790.706313] env[63175]: DEBUG oslo.service.loopingcall [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.706690] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.706887] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26654ec0-9a2a-45de-8c10-bd35ba1cc733 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.726301] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.726301] env[63175]: value = "task-1247999" [ 790.726301] env[63175]: _type = "Task" [ 790.726301] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.733748] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247999, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.764329] env[63175]: DEBUG nova.compute.manager [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Received event network-changed-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 790.764539] env[63175]: DEBUG nova.compute.manager [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Refreshing instance network info cache due to event network-changed-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 790.764719] env[63175]: DEBUG oslo_concurrency.lockutils [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] Acquiring lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.765463] env[63175]: DEBUG oslo_concurrency.lockutils [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] Acquired lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.765463] env[63175]: DEBUG nova.network.neutron [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Refreshing network info cache for port 5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.841767] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 790.865801] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.866063] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.866227] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.866415] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.866560] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.866701] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.866898] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.867070] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.867382] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.867382] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.867553] env[63175]: DEBUG nova.virt.hardware [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.868388] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7efe3ec-ff9b-4759-9bec-ef328a6ff15c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.875756] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebc5250-9f4c-476b-862f-3c33f8b03d58 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.975010] env[63175]: INFO nova.compute.manager [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Took 32.78 seconds to build instance. [ 791.020058] env[63175]: DEBUG oslo_vmware.api [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Task: {'id': task-1247995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166065} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.020267] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 791.020447] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 791.020620] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.020790] env[63175]: INFO nova.compute.manager [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 791.021029] env[63175]: DEBUG oslo.service.loopingcall [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.021218] env[63175]: DEBUG nova.compute.manager [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 791.021311] env[63175]: DEBUG nova.network.neutron [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.128123] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247996, 'name': Rename_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.167069] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Successfully updated port: 595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.211094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.211094] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 791.212135] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.189s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.214457] env[63175]: INFO nova.compute.claims [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.240398] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247999, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.411434] env[63175]: DEBUG nova.compute.manager [req-36fdddf5-7b15-4b41-a702-814b61bd3d82 req-d866869b-7e06-458f-9430-6edf2a6b5175 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Received event network-vif-deleted-f14124e5-edc9-4afb-9633-b5e7d54b3921 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 791.411917] env[63175]: INFO nova.compute.manager [req-36fdddf5-7b15-4b41-a702-814b61bd3d82 req-d866869b-7e06-458f-9430-6edf2a6b5175 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Neutron deleted interface f14124e5-edc9-4afb-9633-b5e7d54b3921; detaching it from the instance and deleting it from the info cache [ 791.412194] env[63175]: DEBUG nova.network.neutron [req-36fdddf5-7b15-4b41-a702-814b61bd3d82 req-d866869b-7e06-458f-9430-6edf2a6b5175 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.476977] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28cf3280-a837-4848-9659-bce6ce792e3f tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.912s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.544735] env[63175]: DEBUG nova.network.neutron [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Updated VIF entry in instance network info cache for port 5301b2d9-86e8-4b5b-abd3-6443a9d5be1a. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 791.545209] env[63175]: DEBUG nova.network.neutron [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Updating instance_info_cache with network_info: [{"id": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "address": "fa:16:3e:60:57:ff", "network": {"id": "174eebb3-caef-4b69-89e0-15cf23f34574", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2090229537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1847b48d1b484fdcbfc114fcf0c4b5e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5301b2d9-86", "ovs_interfaceid": "5301b2d9-86e8-4b5b-abd3-6443a9d5be1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.626819] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1247996, 'name': Rename_Task, 'duration_secs': 0.735816} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.627138] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.627365] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2980d1f4-9632-41e4-8f81-eb9562b463b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.633153] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 791.633153] env[63175]: value = "task-1248000" [ 791.633153] env[63175]: _type = "Task" [ 791.633153] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.640073] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.670931] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.671147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.671338] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.719767] env[63175]: DEBUG nova.compute.utils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.723210] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 791.723415] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.737793] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1247999, 'name': CreateVM_Task, 'duration_secs': 0.783678} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.737950] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.738612] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.738772] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.739094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.739328] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a01dec37-2161-44da-b1c7-80f0bedd66d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.743975] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 791.743975] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5216d7fc-032b-c51b-4ea0-a9bdc7cbca0e" [ 791.743975] env[63175]: _type = "Task" [ 791.743975] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.752028] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5216d7fc-032b-c51b-4ea0-a9bdc7cbca0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.765661] env[63175]: DEBUG nova.policy [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14a8b3d2f6574e37b3125aba432bc56a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e8826d0f0384b56aa862756787ad017', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.823335] env[63175]: DEBUG nova.network.neutron [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.917262] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e82de70d-bdd2-4f63-9346-e1f77cf558b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.926235] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5711745-b433-4483-a8f1-1569aca0cd63 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.952187] env[63175]: DEBUG nova.compute.manager [req-36fdddf5-7b15-4b41-a702-814b61bd3d82 req-d866869b-7e06-458f-9430-6edf2a6b5175 service nova] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Detach interface failed, port_id=f14124e5-edc9-4afb-9633-b5e7d54b3921, reason: Instance 39e55060-73a7-4dbd-96cf-bc48d8737c1c could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 791.979659] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 792.047880] env[63175]: DEBUG oslo_concurrency.lockutils [req-40e43b70-54e3-4c9f-9294-3778865e6039 req-79b9d63a-552b-46f8-9a93-a01e8fa66f1b service nova] Releasing lock "refresh_cache-a2e70590-17ed-4804-b232-57526e87d22b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.087558] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Successfully created port: 71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.143046] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248000, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.201442] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.227014] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 792.263632] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5216d7fc-032b-c51b-4ea0-a9bdc7cbca0e, 'name': SearchDatastore_Task, 'duration_secs': 0.009151} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.264622] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.265498] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.265498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.265498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.265498] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.265888] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98befbdf-3f23-4bd0-96e6-7dc59125e5bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.273953] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.274158] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 792.274851] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e9b5d11-e37f-459d-a3bc-14799a9cb278 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.279547] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 792.279547] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52edc30e-22d0-df66-319b-5b4c85a9084c" [ 792.279547] env[63175]: _type = "Task" [ 792.279547] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.286742] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52edc30e-22d0-df66-319b-5b4c85a9084c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.327464] env[63175]: INFO nova.compute.manager [-] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Took 1.31 seconds to deallocate network for instance. [ 792.371587] env[63175]: DEBUG nova.network.neutron [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Updating instance_info_cache with network_info: [{"id": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "address": "fa:16:3e:5e:9c:11", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap595cdfbb-7c", "ovs_interfaceid": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.458713] env[63175]: DEBUG nova.compute.manager [None req-a5dea9bf-c6cb-456c-bb39-a1bdcf273536 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 792.459799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835a98cc-e98f-4e09-bb9d-97f4c7c8098c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.499912] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.512047] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1f1777-2e7d-435c-a894-0d5fd5f59c1e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.519588] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3796fc0-6870-41b6-bd13-626603c4a1b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.522964] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.523208] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.523409] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.523586] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.523753] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.526737] env[63175]: INFO nova.compute.manager [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Terminating instance [ 792.555432] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4653c15c-7306-4e50-bce0-284d0cebee00 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.560839] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa51b499-56a8-4295-b9d9-e2a62e14fd3c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.576253] env[63175]: DEBUG nova.compute.provider_tree [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.644101] env[63175]: DEBUG oslo_vmware.api [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248000, 'name': PowerOnVM_Task, 'duration_secs': 0.517704} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.644393] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.644573] env[63175]: INFO nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Took 9.06 seconds to spawn the instance on the hypervisor. [ 792.644751] env[63175]: DEBUG nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 792.645530] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefc1771-f731-4e4a-89b2-39dc00e9fbc5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.789782] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52edc30e-22d0-df66-319b-5b4c85a9084c, 'name': SearchDatastore_Task, 'duration_secs': 0.007951} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.790700] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5c17631-537e-4404-ac9e-dbaf6c7b138c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.795703] env[63175]: DEBUG nova.compute.manager [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Received event network-vif-plugged-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 792.795981] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Acquiring lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.796255] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.796514] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.796750] env[63175]: DEBUG nova.compute.manager [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] No waiting events found dispatching network-vif-plugged-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 792.796959] env[63175]: WARNING nova.compute.manager [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Received unexpected event network-vif-plugged-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f for instance with vm_state building and task_state spawning. [ 792.797182] env[63175]: DEBUG nova.compute.manager [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Received event network-changed-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 792.797352] env[63175]: DEBUG nova.compute.manager [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Refreshing instance network info cache due to event network-changed-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 792.797518] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Acquiring lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.801039] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 792.801039] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8e4f9-f683-d02f-b2b3-ece031ba0b7f" [ 792.801039] env[63175]: _type = "Task" [ 792.801039] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.810633] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8e4f9-f683-d02f-b2b3-ece031ba0b7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.835912] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.873133] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.873589] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance network_info: |[{"id": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "address": "fa:16:3e:5e:9c:11", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap595cdfbb-7c", "ovs_interfaceid": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 792.873987] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Acquired lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.874282] env[63175]: DEBUG nova.network.neutron [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Refreshing network info cache for port 595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.875624] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:9c:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.882957] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating folder: Project (74dd39e20aad4ddb874282973cea02bc). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.883481] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4542e899-053f-4347-95c3-e9ee803f41a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.894626] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created folder: Project (74dd39e20aad4ddb874282973cea02bc) in parent group-v268956. [ 792.894820] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating folder: Instances. Parent ref: group-v269002. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.895068] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dccb9ddd-c739-4942-b966-f4e88a9808eb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.903897] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created folder: Instances in parent group-v269002. [ 792.904180] env[63175]: DEBUG oslo.service.loopingcall [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.904349] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.904551] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-364b28cb-0aca-496f-b564-93f328e0efba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.922641] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.922641] env[63175]: value = "task-1248003" [ 792.922641] env[63175]: _type = "Task" [ 792.922641] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.931673] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248003, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.971354] env[63175]: INFO nova.compute.manager [None req-a5dea9bf-c6cb-456c-bb39-a1bdcf273536 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] instance snapshotting [ 792.972052] env[63175]: DEBUG nova.objects.instance [None req-a5dea9bf-c6cb-456c-bb39-a1bdcf273536 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lazy-loading 'flavor' on Instance uuid 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.057019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "refresh_cache-35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.057342] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquired lock "refresh_cache-35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.057620] env[63175]: DEBUG nova.network.neutron [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.079355] env[63175]: DEBUG nova.scheduler.client.report [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 793.165736] env[63175]: INFO nova.compute.manager [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Took 37.27 seconds to build instance. [ 793.239978] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 793.291530] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.291792] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.292013] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.292271] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.292449] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.292606] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.292815] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.292984] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.293172] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.293393] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.293615] env[63175]: DEBUG nova.virt.hardware [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.294567] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4918ef62-6ac4-427c-ba27-c00f0b790c33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.301997] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78298ad-daf4-4fad-b021-4ce5e410bcf6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.313957] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8e4f9-f683-d02f-b2b3-ece031ba0b7f, 'name': SearchDatastore_Task, 'duration_secs': 0.009583} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.321489] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.321753] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a2e70590-17ed-4804-b232-57526e87d22b/a2e70590-17ed-4804-b232-57526e87d22b.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.322549] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fd38f08-2ccc-4d0e-8f26-99ca9a8093e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.328261] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 793.328261] env[63175]: value = "task-1248004" [ 793.328261] env[63175]: _type = "Task" [ 793.328261] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.335732] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.432934] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248003, 'name': CreateVM_Task, 'duration_secs': 0.414485} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.433011] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.434605] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.434772] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.435320] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.435606] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11e9ac1a-3da5-4095-8839-d87edff8d1f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.441923] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 793.441923] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d2313-29a8-056f-bd73-b888c1910cd5" [ 793.441923] env[63175]: _type = "Task" [ 793.441923] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.450206] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d2313-29a8-056f-bd73-b888c1910cd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.478710] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01525c8e-5ab7-4426-99c8-e3cbfc288d99 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.501044] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8455834-0f36-47b3-818d-43260cd651d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.583956] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.584899] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 793.587972] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.461s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.591021] env[63175]: INFO nova.compute.claims [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.593760] env[63175]: DEBUG nova.network.neutron [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.667881] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c49ac142-6205-4d19-be6b-1a3593c73fe6 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.555s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.703316] env[63175]: DEBUG nova.network.neutron [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.714594] env[63175]: DEBUG nova.network.neutron [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Updated VIF entry in instance network info cache for port 595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 793.715107] env[63175]: DEBUG nova.network.neutron [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Updating instance_info_cache with network_info: [{"id": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "address": "fa:16:3e:5e:9c:11", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap595cdfbb-7c", "ovs_interfaceid": "595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.739803] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Successfully updated port: 71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.842090] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442161} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.842328] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a2e70590-17ed-4804-b232-57526e87d22b/a2e70590-17ed-4804-b232-57526e87d22b.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.842553] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.842809] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4995edf6-f2e9-4fa5-b54a-1c38b3be5765 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.849759] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 793.849759] env[63175]: value = "task-1248005" [ 793.849759] env[63175]: _type = "Task" [ 793.849759] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.857603] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248005, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.952154] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d2313-29a8-056f-bd73-b888c1910cd5, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.952683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.952683] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.952830] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.953058] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.953423] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.953714] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45c34e97-e88f-45dc-af36-cf88519f19ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.961729] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.961934] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.962739] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f972f1db-c692-4548-ac9b-bd7fae057da9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.968468] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 793.968468] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522c0d33-e402-0169-93e2-984a56d8a5fe" [ 793.968468] env[63175]: _type = "Task" [ 793.968468] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.975854] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522c0d33-e402-0169-93e2-984a56d8a5fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.012093] env[63175]: DEBUG nova.compute.manager [None req-a5dea9bf-c6cb-456c-bb39-a1bdcf273536 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance disappeared during snapshot {{(pid=63175) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 794.096066] env[63175]: DEBUG nova.compute.utils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.097450] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 794.097610] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.146054] env[63175]: DEBUG nova.compute.manager [None req-a5dea9bf-c6cb-456c-bb39-a1bdcf273536 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Found 0 images (rotation: 2) {{(pid=63175) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 794.173271] env[63175]: DEBUG nova.policy [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1225f2968ab04a5baf6af02f95314a6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f27d5fdc63dc4d5b96b85b55bc8b428d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.174863] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 794.206587] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Releasing lock "refresh_cache-35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.207045] env[63175]: DEBUG nova.compute.manager [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 794.207253] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.208295] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4241f2-3f84-46c3-8676-f8f82c0439a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.216873] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.217906] env[63175]: DEBUG oslo_concurrency.lockutils [req-b78c6392-7bc4-43e1-b4ee-adccf4f15a7a req-cba8cc5e-c3d5-4bad-847e-f79f6a8ab133 service nova] Releasing lock "refresh_cache-f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.218276] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38d4b1cf-d0ea-451e-bec8-1c6b62a314ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.224166] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 794.224166] env[63175]: value = "task-1248006" [ 794.224166] env[63175]: _type = "Task" [ 794.224166] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.232299] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1248006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.243323] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.243464] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.243617] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.359242] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248005, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.185957} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.359642] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.360444] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54729ccd-7c6b-4046-a520-7f866c35c1b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.382550] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] a2e70590-17ed-4804-b232-57526e87d22b/a2e70590-17ed-4804-b232-57526e87d22b.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.383197] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2c8df12-3359-48c9-a29d-6aa4304891d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.402646] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 794.402646] env[63175]: value = "task-1248007" [ 794.402646] env[63175]: _type = "Task" [ 794.402646] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.412782] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248007, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.478806] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522c0d33-e402-0169-93e2-984a56d8a5fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009362} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.479617] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb537beb-10c5-4494-b96c-1e27ff029e22 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.486733] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 794.486733] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5223445b-caf9-8ab2-e50c-eb54fb22b67b" [ 794.486733] env[63175]: _type = "Task" [ 794.486733] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.494432] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5223445b-caf9-8ab2-e50c-eb54fb22b67b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.602331] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 794.699482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.738418] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1248006, 'name': PowerOffVM_Task, 'duration_secs': 0.129421} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.742289] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.742900] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 794.743320] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-556147ce-92ca-45a3-a17d-f1eb47f155f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.769487] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.769487] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.769487] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Deleting the datastore file [datastore2] 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.769662] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af12d923-912f-4e9e-a4da-0fbe68623810 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.781601] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for the task: (returnval){ [ 794.781601] env[63175]: value = "task-1248009" [ 794.781601] env[63175]: _type = "Task" [ 794.781601] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.782394] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.793655] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1248009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.817172] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Successfully created port: 810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.912481] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.975652] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Received event network-vif-plugged-71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 794.975897] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.976530] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.976777] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.976922] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] No waiting events found dispatching network-vif-plugged-71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 794.977085] env[63175]: WARNING nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Received unexpected event network-vif-plugged-71756e11-f67d-4268-a4b0-25b0a8cdefbf for instance with vm_state building and task_state spawning. [ 794.977254] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Received event network-changed-71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 794.977606] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Refreshing instance network info cache due to event network-changed-71756e11-f67d-4268-a4b0-25b0a8cdefbf. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 794.977817] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.997208] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5223445b-caf9-8ab2-e50c-eb54fb22b67b, 'name': SearchDatastore_Task, 'duration_secs': 0.00856} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.997483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.997739] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.998015] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c6cf262-0ecd-442f-9e98-dc30bf290d3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.001383] env[63175]: DEBUG nova.network.neutron [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.010028] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 795.010028] env[63175]: value = "task-1248010" [ 795.010028] env[63175]: _type = "Task" [ 795.010028] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.019187] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.061641] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ba58aa-90bb-4e95-ae36-85f14350d798 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.070628] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc00379b-fcc4-4788-b108-e19b6e32a1be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.112196] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b047d1d7-3bb5-4caa-9772-b549c69b2842 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.119431] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57753c8d-8c97-4213-9838-6574df2e5e02 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.133315] env[63175]: DEBUG nova.compute.provider_tree [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.295503] env[63175]: DEBUG oslo_vmware.api [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Task: {'id': task-1248009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114908} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.295503] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.295863] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 795.295916] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.296156] env[63175]: INFO nova.compute.manager [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Took 1.09 seconds to destroy the instance on the hypervisor. [ 795.296375] env[63175]: DEBUG oslo.service.loopingcall [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.296583] env[63175]: DEBUG nova.compute.manager [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 795.296668] env[63175]: DEBUG nova.network.neutron [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.325827] env[63175]: DEBUG nova.network.neutron [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.413366] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248007, 'name': ReconfigVM_Task, 'duration_secs': 0.539348} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.413708] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Reconfigured VM instance instance-00000038 to attach disk [datastore1] a2e70590-17ed-4804-b232-57526e87d22b/a2e70590-17ed-4804-b232-57526e87d22b.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.414397] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-084623fe-932f-4c2a-ba86-67a33b3ea10d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.421871] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 795.421871] env[63175]: value = "task-1248011" [ 795.421871] env[63175]: _type = "Task" [ 795.421871] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.430786] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248011, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.505013] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.505448] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Instance network_info: |[{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 795.505773] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.505989] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Refreshing network info cache for port 71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.507243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:ea:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '777870ab-362f-4a17-9c1c-8d9cc26cd4ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71756e11-f67d-4268-a4b0-25b0a8cdefbf', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.515325] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Creating folder: Project (0e8826d0f0384b56aa862756787ad017). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.516562] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8860379b-e8a2-4afc-8079-e08da8543742 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.526866] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248010, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510933} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.527109] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.527323] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.527550] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0af7fff3-014d-47ad-a6c0-11d8c43f4f0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.530189] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Created folder: Project (0e8826d0f0384b56aa862756787ad017) in parent group-v268956. [ 795.530366] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Creating folder: Instances. Parent ref: group-v269005. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.530869] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-412cc002-5902-4b43-a996-ccd69c9e579f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.534557] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 795.534557] env[63175]: value = "task-1248013" [ 795.534557] env[63175]: _type = "Task" [ 795.534557] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.542087] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248013, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.543235] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Created folder: Instances in parent group-v269005. [ 795.543457] env[63175]: DEBUG oslo.service.loopingcall [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.543632] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 795.543820] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b5fa89d-7af3-4d26-a814-07a7270b4b91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.563470] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.563470] env[63175]: value = "task-1248015" [ 795.563470] env[63175]: _type = "Task" [ 795.563470] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.570940] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.613054] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 795.636936] env[63175]: DEBUG nova.scheduler.client.report [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 795.642705] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.642925] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.643120] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.643525] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.643525] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.643615] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.643812] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.643970] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.644291] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.644551] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.644802] env[63175]: DEBUG nova.virt.hardware [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.645759] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3922e664-0c46-4385-b97a-b2771d153911 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.654202] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116c037c-a060-4602-ace4-9344e7146604 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.829182] env[63175]: DEBUG nova.network.neutron [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.933884] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248011, 'name': Rename_Task, 'duration_secs': 0.185011} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.933884] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.934066] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adacfbe7-ee42-4e56-8d1a-8e451e016087 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.941106] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 795.941106] env[63175]: value = "task-1248016" [ 795.941106] env[63175]: _type = "Task" [ 795.941106] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.948384] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.045750] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248013, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104989} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.048145] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.049075] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236821a9-88c2-4700-add8-54449a03818d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.076599] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.079867] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1371e820-874f-494a-9bbc-e9089c0b2ed3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.102326] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.104028] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 796.104028] env[63175]: value = "task-1248017" [ 796.104028] env[63175]: _type = "Task" [ 796.104028] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.111822] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.149820] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.150542] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 796.153753] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.302s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.153879] env[63175]: DEBUG nova.objects.instance [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lazy-loading 'resources' on Instance uuid 12a2d788-c995-468b-b2a2-17dba8ad01db {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.270650] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updated VIF entry in instance network info cache for port 71756e11-f67d-4268-a4b0-25b0a8cdefbf. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 796.270997] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.335948] env[63175]: INFO nova.compute.manager [-] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Took 1.04 seconds to deallocate network for instance. [ 796.451783] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248016, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.568138] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Successfully updated port: 810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.578264] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.615440] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.646363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.646363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.646363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.646363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.646519] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.648357] env[63175]: INFO nova.compute.manager [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Terminating instance [ 796.657082] env[63175]: DEBUG nova.compute.utils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.662066] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 796.662066] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.700809] env[63175]: DEBUG nova.policy [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db76f4c61f9a4aea80c546253b937b3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7126014d48214eda98fb43c1b774133f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.775614] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.775920] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-changed-037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 796.776171] env[63175]: DEBUG nova.compute.manager [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing instance network info cache due to event network-changed-037fa23d-fc83-4fdc-810b-b97533910bbe. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 796.776443] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.776608] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.776803] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing network info cache for port 037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.846008] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.940845] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde33095-0286-4f04-8637-6ef3fd72f8d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.952391] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248016, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.955153] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478e3304-a09d-4cc1-a1f2-dd9116e741e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.985344] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b5d5eb-e119-44ba-a032-da9fbe1b3d7e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.993352] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8466a7-029e-4edb-8405-4e1006d15cc7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.007084] env[63175]: DEBUG nova.compute.provider_tree [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.010030] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Successfully created port: ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.012666] env[63175]: DEBUG nova.compute.manager [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Received event network-vif-plugged-810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 797.012909] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Acquiring lock "72d131a0-2617-49a1-8aff-897908929bb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.013122] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Lock "72d131a0-2617-49a1-8aff-897908929bb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.013291] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Lock "72d131a0-2617-49a1-8aff-897908929bb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.013455] env[63175]: DEBUG nova.compute.manager [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] No waiting events found dispatching network-vif-plugged-810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 797.013617] env[63175]: WARNING nova.compute.manager [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Received unexpected event network-vif-plugged-810e244a-4573-4ae6-98d1-debc881e4ab2 for instance with vm_state building and task_state spawning. [ 797.013775] env[63175]: DEBUG nova.compute.manager [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Received event network-changed-810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 797.013933] env[63175]: DEBUG nova.compute.manager [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Refreshing instance network info cache due to event network-changed-810e244a-4573-4ae6-98d1-debc881e4ab2. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 797.014147] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Acquiring lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.014328] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Acquired lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.014422] env[63175]: DEBUG nova.network.neutron [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Refreshing network info cache for port 810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 797.077285] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.077652] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.115991] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.152781] env[63175]: DEBUG nova.compute.manager [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 797.153754] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.153754] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56d6163a-cf8f-4cc0-ad97-8b0d8c03ad37 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.161948] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 797.161948] env[63175]: value = "task-1248018" [ 797.161948] env[63175]: _type = "Task" [ 797.161948] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.162496] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 797.173516] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.454776] env[63175]: DEBUG oslo_vmware.api [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248016, 'name': PowerOnVM_Task, 'duration_secs': 1.48823} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.455376] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 797.455747] env[63175]: INFO nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Took 8.98 seconds to spawn the instance on the hypervisor. [ 797.456093] env[63175]: DEBUG nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 797.456999] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203c21ad-7ad2-46d4-a1bc-1d185e88d6a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.516663] env[63175]: DEBUG nova.scheduler.client.report [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 797.571929] env[63175]: DEBUG nova.network.neutron [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.585163] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.615599] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248017, 'name': ReconfigVM_Task, 'duration_secs': 1.086695} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.615938] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Reconfigured VM instance instance-00000039 to attach disk [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.616590] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9fcaba0-8db7-46d3-84fa-7a109e89068f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.622780] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 797.622780] env[63175]: value = "task-1248019" [ 797.622780] env[63175]: _type = "Task" [ 797.622780] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.632439] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248019, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.639919] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updated VIF entry in instance network info cache for port 037fa23d-fc83-4fdc-810b-b97533910bbe. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.639919] env[63175]: DEBUG nova.network.neutron [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.677593] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248018, 'name': PowerOffVM_Task, 'duration_secs': 0.180397} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.677864] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.678076] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 797.678273] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268972', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'name': 'volume-ac967057-2445-49c8-8181-4d0115179902', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e137904f-dc43-4ebb-90ab-e10ea5487fe5', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'serial': 'ac967057-2445-49c8-8181-4d0115179902'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 797.679080] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086cb12d-4c8c-4b20-9285-04803fb1c664 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.699866] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da123d69-20a4-48f9-9bb4-bff555688319 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.704078] env[63175]: DEBUG nova.network.neutron [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.711035] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384d6271-14d9-4fa5-ad7d-66d2733f8477 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.729179] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa215e44-2e40-40cd-a49c-63db367a060c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.745492] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] The volume has not been displaced from its original location: [datastore2] volume-ac967057-2445-49c8-8181-4d0115179902/volume-ac967057-2445-49c8-8181-4d0115179902.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 797.750719] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Reconfiguring VM instance instance-00000031 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 797.751075] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9098c41c-6882-4289-99e5-9f66c14d9ab9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.770215] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 797.770215] env[63175]: value = "task-1248020" [ 797.770215] env[63175]: _type = "Task" [ 797.770215] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.778805] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248020, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.979489] env[63175]: INFO nova.compute.manager [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Took 37.97 seconds to build instance. [ 798.022834] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.026342] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.270s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.027261] env[63175]: INFO nova.compute.claims [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.046034] env[63175]: INFO nova.scheduler.client.report [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Deleted allocations for instance 12a2d788-c995-468b-b2a2-17dba8ad01db [ 798.078872] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248015, 'name': CreateVM_Task, 'duration_secs': 2.156277} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.079116] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.079763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.079933] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.080338] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.080637] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c74fad-88d6-4cc2-8cd2-b84b3c73b40d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.085699] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 798.085699] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5282d9dd-b01b-b551-706b-73f026b78d36" [ 798.085699] env[63175]: _type = "Task" [ 798.085699] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.094987] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5282d9dd-b01b-b551-706b-73f026b78d36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.132516] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248019, 'name': Rename_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.142224] env[63175]: DEBUG oslo_concurrency.lockutils [req-75c9a818-2075-401f-87a5-daa3b493141e req-3d21439b-2281-4da0-b1e6-f1521299bd46 service nova] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.174087] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 798.200205] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.200478] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.200641] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.200827] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.200976] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.201146] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.201357] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.201519] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.201691] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.201856] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.202252] env[63175]: DEBUG nova.virt.hardware [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.203164] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6014cb-2081-4f75-a9a9-e7601ec81bb0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.206011] env[63175]: DEBUG oslo_concurrency.lockutils [req-a74385cf-64c8-4b9f-8df2-3923293b4f87 req-e5dcf585-2072-4164-a4a6-fd694c6264f6 service nova] Releasing lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.206366] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquired lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.206519] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.213176] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3461cb-49c0-4f37-b85b-8ef50881d432 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.281592] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248020, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.481969] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ef8be359-a6d6-4677-8e71-b67a156d2efb tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.367s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.552685] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4fd69449-08f2-4774-8158-2db6adfae650 tempest-ServersAdminTestJSON-205669498 tempest-ServersAdminTestJSON-205669498-project-member] Lock "12a2d788-c995-468b-b2a2-17dba8ad01db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.229s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.597116] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5282d9dd-b01b-b551-706b-73f026b78d36, 'name': SearchDatastore_Task, 'duration_secs': 0.009994} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.597116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.597116] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.597116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.597381] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.597381] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.597582] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6e14e21-8149-4175-b345-9569e0157a40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.606465] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.606651] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.607449] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a02d48e-ea76-401c-917b-2586e5817570 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.612837] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 798.612837] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf1dc0-762f-5def-7981-b8e4905605ca" [ 798.612837] env[63175]: _type = "Task" [ 798.612837] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.620562] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf1dc0-762f-5def-7981-b8e4905605ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.631274] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248019, 'name': Rename_Task, 'duration_secs': 0.994048} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.631524] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 798.631859] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb459733-6d50-4462-9350-9691ac05d852 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.638554] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 798.638554] env[63175]: value = "task-1248021" [ 798.638554] env[63175]: _type = "Task" [ 798.638554] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.646627] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.738079] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.782860] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248020, 'name': ReconfigVM_Task, 'duration_secs': 0.857935} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.783179] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Reconfigured VM instance instance-00000031 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 798.788626] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e309ac8b-5445-458c-90df-747802e78b8f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.803981] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 798.803981] env[63175]: value = "task-1248022" [ 798.803981] env[63175]: _type = "Task" [ 798.803981] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.818216] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248022, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.883668] env[63175]: DEBUG nova.network.neutron [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Updating instance_info_cache with network_info: [{"id": "810e244a-4573-4ae6-98d1-debc881e4ab2", "address": "fa:16:3e:83:34:68", "network": {"id": "8c2a8bf0-07fd-4f3d-b486-c1cb5e8a6459", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1266026319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f27d5fdc63dc4d5b96b85b55bc8b428d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap810e244a-45", "ovs_interfaceid": "810e244a-4573-4ae6-98d1-debc881e4ab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.915084] env[63175]: DEBUG oslo_concurrency.lockutils [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "interface-a2e70590-17ed-4804-b232-57526e87d22b-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.915350] env[63175]: DEBUG oslo_concurrency.lockutils [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "interface-a2e70590-17ed-4804-b232-57526e87d22b-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.915862] env[63175]: DEBUG nova.objects.instance [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lazy-loading 'flavor' on Instance uuid a2e70590-17ed-4804-b232-57526e87d22b {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.944061] env[63175]: DEBUG nova.compute.manager [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Received event network-vif-plugged-ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 798.944061] env[63175]: DEBUG oslo_concurrency.lockutils [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] Acquiring lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.944291] env[63175]: DEBUG oslo_concurrency.lockutils [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] Lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.944413] env[63175]: DEBUG oslo_concurrency.lockutils [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] Lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.944581] env[63175]: DEBUG nova.compute.manager [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] No waiting events found dispatching network-vif-plugged-ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 798.944741] env[63175]: WARNING nova.compute.manager [req-f2c76602-c3d0-4b7e-96cd-db201b9527d2 req-f59fa6b1-7811-42a2-a1bb-294f312d3663 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Received unexpected event network-vif-plugged-ef671ac5-130d-4e3e-a8e6-2650bbf1c278 for instance with vm_state building and task_state spawning. [ 798.984650] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 799.117675] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Successfully updated port: ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.135967] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf1dc0-762f-5def-7981-b8e4905605ca, 'name': SearchDatastore_Task, 'duration_secs': 0.008956} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.136870] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77727e63-67c7-4a9b-952a-4ab877462765 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.148082] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 799.148082] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52490b31-f0ae-f5a9-7251-fc51838e9b06" [ 799.148082] env[63175]: _type = "Task" [ 799.148082] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.155326] env[63175]: DEBUG oslo_vmware.api [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248021, 'name': PowerOnVM_Task, 'duration_secs': 0.476311} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.156787] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.157106] env[63175]: INFO nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Took 8.32 seconds to spawn the instance on the hypervisor. [ 799.157372] env[63175]: DEBUG nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 799.160519] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9426d5-1d00-49e8-8d47-bfcbdc1ca201 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.167392] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52490b31-f0ae-f5a9-7251-fc51838e9b06, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.168178] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.168486] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3/cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.168789] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-664165c7-7338-4579-8e84-550738abf092 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.180918] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 799.180918] env[63175]: value = "task-1248023" [ 799.180918] env[63175]: _type = "Task" [ 799.180918] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.189272] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.316280] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248022, 'name': ReconfigVM_Task, 'duration_secs': 0.117206} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.316894] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268972', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'name': 'volume-ac967057-2445-49c8-8181-4d0115179902', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e137904f-dc43-4ebb-90ab-e10ea5487fe5', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac967057-2445-49c8-8181-4d0115179902', 'serial': 'ac967057-2445-49c8-8181-4d0115179902'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 799.317413] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.318414] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b002f497-616e-47f3-a773-19dc369f060a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.324804] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da9cf84-642a-431d-83ee-641602b10d80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.329961] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 799.330701] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23c3fb8c-87a8-40a6-a154-64e45dfeebaa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.335415] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35f993c-364a-433c-a306-a08a7c241ed7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.371348] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babc3b6c-80bb-43f0-ae92-c5289027e9f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.379911] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dbbdb09-f1c3-407a-b9c6-841cf6e646d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.386573] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Releasing lock "refresh_cache-72d131a0-2617-49a1-8aff-897908929bb0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.386957] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance network_info: |[{"id": "810e244a-4573-4ae6-98d1-debc881e4ab2", "address": "fa:16:3e:83:34:68", "network": {"id": "8c2a8bf0-07fd-4f3d-b486-c1cb5e8a6459", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1266026319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f27d5fdc63dc4d5b96b85b55bc8b428d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap810e244a-45", "ovs_interfaceid": "810e244a-4573-4ae6-98d1-debc881e4ab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 799.387522] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:34:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '810e244a-4573-4ae6-98d1-debc881e4ab2', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.394838] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Creating folder: Project (f27d5fdc63dc4d5b96b85b55bc8b428d). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.403701] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-498fb01f-2220-48ae-b216-57f625178e42 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.405598] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 799.405807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 799.405987] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Deleting the datastore file [datastore2] e137904f-dc43-4ebb-90ab-e10ea5487fe5 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 799.406502] env[63175]: DEBUG nova.compute.provider_tree [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.407699] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a4d659f-dc32-4393-8d50-38151bdf005c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.414523] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for the task: (returnval){ [ 799.414523] env[63175]: value = "task-1248026" [ 799.414523] env[63175]: _type = "Task" [ 799.414523] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.420085] env[63175]: DEBUG nova.objects.instance [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lazy-loading 'pci_requests' on Instance uuid a2e70590-17ed-4804-b232-57526e87d22b {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.421155] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Created folder: Project (f27d5fdc63dc4d5b96b85b55bc8b428d) in parent group-v268956. [ 799.421414] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Creating folder: Instances. Parent ref: group-v269008. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.422653] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70ef6a6d-783d-482c-ab55-2a8e835bdb24 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.429528] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.432577] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Created folder: Instances in parent group-v269008. [ 799.432577] env[63175]: DEBUG oslo.service.loopingcall [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.433264] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.433407] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3be309fe-1d3e-4e15-bde2-2676c610e7f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.453075] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.453075] env[63175]: value = "task-1248028" [ 799.453075] env[63175]: _type = "Task" [ 799.453075] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.461956] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248028, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.507241] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.629044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.629044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.629044] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.688830] env[63175]: INFO nova.compute.manager [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Took 37.56 seconds to build instance. [ 799.693507] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442016} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.694037] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3/cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.694037] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.694210] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f04db0e-d0ff-4737-82b2-6e7243575f0e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.700570] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 799.700570] env[63175]: value = "task-1248029" [ 799.700570] env[63175]: _type = "Task" [ 799.700570] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.710842] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248029, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.906252] env[63175]: DEBUG nova.compute.manager [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Received event network-changed-ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 799.906252] env[63175]: DEBUG nova.compute.manager [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Refreshing instance network info cache due to event network-changed-ef671ac5-130d-4e3e-a8e6-2650bbf1c278. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 799.906252] env[63175]: DEBUG oslo_concurrency.lockutils [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] Acquiring lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.913023] env[63175]: DEBUG nova.scheduler.client.report [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 799.923451] env[63175]: DEBUG nova.objects.base [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 799.923988] env[63175]: DEBUG nova.network.neutron [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.930657] env[63175]: DEBUG oslo_vmware.api [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Task: {'id': task-1248026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211062} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.932134] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.932496] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.932789] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.933776] env[63175]: INFO nova.compute.manager [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Took 2.78 seconds to destroy the instance on the hypervisor. [ 799.934201] env[63175]: DEBUG oslo.service.loopingcall [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.937892] env[63175]: DEBUG nova.compute.manager [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 799.937892] env[63175]: DEBUG nova.network.neutron [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.962860] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248028, 'name': CreateVM_Task, 'duration_secs': 0.357015} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.963058] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.963740] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.963994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.964665] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.964896] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34da2c22-dc75-4c1d-9a33-0dcacaee0bb5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.970374] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 799.970374] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52929115-0ddf-7238-f77a-3631421ce4ef" [ 799.970374] env[63175]: _type = "Task" [ 799.970374] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.979201] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52929115-0ddf-7238-f77a-3631421ce4ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.141965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-db335398-6cf3-46de-a791-60f1d4995f6c tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "interface-a2e70590-17ed-4804-b232-57526e87d22b-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.226s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.190821] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df759f-bcdd-4220-a628-fde841b690b8 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.888s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.207123] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.219525] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248029, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062978} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.219827] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.221707] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6275aeca-a552-4811-a2e5-bb7ed237d445 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.246714] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3/cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.249987] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb48eac3-524a-4d2a-a0cb-0477164a6c95 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.276878] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 800.276878] env[63175]: value = "task-1248030" [ 800.276878] env[63175]: _type = "Task" [ 800.276878] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.286238] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248030, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.417906] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.419304] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 800.421942] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.669s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.424557] env[63175]: INFO nova.compute.claims [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.472214] env[63175]: DEBUG nova.network.neutron [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Updating instance_info_cache with network_info: [{"id": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "address": "fa:16:3e:21:0e:5a", "network": {"id": "45f31408-e9ed-44de-9b57-3303a8d7c2b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-725769398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7126014d48214eda98fb43c1b774133f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef671ac5-13", "ovs_interfaceid": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.488435] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52929115-0ddf-7238-f77a-3631421ce4ef, 'name': SearchDatastore_Task, 'duration_secs': 0.009135} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.492586] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.492586] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.492586] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.492586] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.492766] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.492766] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6236bc34-7453-4d58-a620-2c7b0ec131f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.502225] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.504980] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 800.504980] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa4a4471-033a-4170-915d-892c94899063 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.511485] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 800.511485] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525ca678-b050-aa55-b6cf-3779cb9dc979" [ 800.511485] env[63175]: _type = "Task" [ 800.511485] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.520511] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525ca678-b050-aa55-b6cf-3779cb9dc979, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.693605] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 800.788529] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248030, 'name': ReconfigVM_Task, 'duration_secs': 0.29804} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.788529] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfigured VM instance instance-0000003a to attach disk [datastore1] cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3/cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.788529] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c573975f-d1f2-4950-b0c7-db003040b1f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.794932] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 800.794932] env[63175]: value = "task-1248031" [ 800.794932] env[63175]: _type = "Task" [ 800.794932] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.804509] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248031, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.935746] env[63175]: DEBUG nova.compute.utils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.937038] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 800.937217] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.973693] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Releasing lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.974172] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Instance network_info: |[{"id": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "address": "fa:16:3e:21:0e:5a", "network": {"id": "45f31408-e9ed-44de-9b57-3303a8d7c2b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-725769398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7126014d48214eda98fb43c1b774133f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef671ac5-13", "ovs_interfaceid": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 800.974572] env[63175]: DEBUG oslo_concurrency.lockutils [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] Acquired lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.974824] env[63175]: DEBUG nova.network.neutron [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Refreshing network info cache for port ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.976717] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:0e:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef671ac5-130d-4e3e-a8e6-2650bbf1c278', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.989644] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Creating folder: Project (7126014d48214eda98fb43c1b774133f). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.993803] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-905a0a17-fda4-45ba-b31a-ee7870d209ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.004137] env[63175]: DEBUG nova.policy [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '564ff9a1de5f43f4b3b05f3592bfa833', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54aee75c86d543a995a364ed78426ec2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.010471] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Created folder: Project (7126014d48214eda98fb43c1b774133f) in parent group-v268956. [ 801.010737] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Creating folder: Instances. Parent ref: group-v269011. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.011796] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-365154a5-858f-45da-a662-dbbc4d94b3a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.030617] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525ca678-b050-aa55-b6cf-3779cb9dc979, 'name': SearchDatastore_Task, 'duration_secs': 0.012665} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.033407] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Created folder: Instances in parent group-v269011. [ 801.033762] env[63175]: DEBUG oslo.service.loopingcall [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.034076] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c31f27ed-885c-4370-a7d9-bf483f73c9bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.037933] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.041456] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7bf4e04-2080-426f-acea-b2a00a82943d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.063141] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 801.063141] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e99534-cb06-f212-b69e-c836fab944ed" [ 801.063141] env[63175]: _type = "Task" [ 801.063141] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.067403] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.067403] env[63175]: value = "task-1248034" [ 801.067403] env[63175]: _type = "Task" [ 801.067403] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.079226] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e99534-cb06-f212-b69e-c836fab944ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.080447] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248034, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.220250] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.258043] env[63175]: DEBUG nova.network.neutron [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.305606] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248031, 'name': Rename_Task, 'duration_secs': 0.157612} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.305905] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 801.306178] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12e881f1-96ef-411e-97f4-b6601580444e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.312557] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 801.312557] env[63175]: value = "task-1248035" [ 801.312557] env[63175]: _type = "Task" [ 801.312557] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.320868] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.329434] env[63175]: DEBUG nova.network.neutron [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Updated VIF entry in instance network info cache for port ef671ac5-130d-4e3e-a8e6-2650bbf1c278. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.329783] env[63175]: DEBUG nova.network.neutron [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Updating instance_info_cache with network_info: [{"id": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "address": "fa:16:3e:21:0e:5a", "network": {"id": "45f31408-e9ed-44de-9b57-3303a8d7c2b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-725769398-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7126014d48214eda98fb43c1b774133f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8fedd232-bfc1-4e7f-bd5e-c43ef8f2f08a", "external-id": "nsx-vlan-transportzone-925", "segmentation_id": 925, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef671ac5-13", "ovs_interfaceid": "ef671ac5-130d-4e3e-a8e6-2650bbf1c278", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.441842] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 801.587640] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e99534-cb06-f212-b69e-c836fab944ed, 'name': SearchDatastore_Task, 'duration_secs': 0.026156} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.591822] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.592207] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 72d131a0-2617-49a1-8aff-897908929bb0/72d131a0-2617-49a1-8aff-897908929bb0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 801.592485] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248034, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.592682] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-063c5c02-1d4b-4775-a81a-150fcfb77a54 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.601311] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 801.601311] env[63175]: value = "task-1248036" [ 801.601311] env[63175]: _type = "Task" [ 801.601311] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.614974] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.761890] env[63175]: INFO nova.compute.manager [-] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Took 1.82 seconds to deallocate network for instance. [ 801.774753] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Successfully created port: 57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.777356] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de85ddb-6664-4750-836f-b81f1f3404c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.786974] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad4539c-6df0-4877-96b2-07314ffddac0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.836770] env[63175]: DEBUG oslo_concurrency.lockutils [req-889be2b7-58df-4277-8697-30f71283944a req-f91a0231-d8f3-4894-972b-c53e1d65f49c service nova] Releasing lock "refresh_cache-4d07a537-7df6-4659-8760-bf7e7925da25" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.842116] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be69d69-d556-44ee-9170-b2727fac90a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.857078] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0819432-5e65-4969-9103-84be25aff448 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.864075] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248035, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.878219] env[63175]: DEBUG nova.compute.provider_tree [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.084038] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248034, 'name': CreateVM_Task, 'duration_secs': 0.554114} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.084224] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.085255] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.085492] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.086110] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.088220] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07f8d315-1ff4-4216-a0dd-4d5db19c69a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.093347] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 802.093347] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527d425f-0b96-ce4b-44bb-1c51fbcbeecc" [ 802.093347] env[63175]: _type = "Task" [ 802.093347] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.100776] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527d425f-0b96-ce4b-44bb-1c51fbcbeecc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.111511] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248036, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496878} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.111764] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 72d131a0-2617-49a1-8aff-897908929bb0/72d131a0-2617-49a1-8aff-897908929bb0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.112058] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.112453] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd6eddff-112b-4441-9389-17233b7920d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.118102] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 802.118102] env[63175]: value = "task-1248037" [ 802.118102] env[63175]: _type = "Task" [ 802.118102] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.125863] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248037, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.206971] env[63175]: DEBUG nova.compute.manager [req-af2499d5-07a5-4f2c-84e3-9c968dd969e0 req-4ff384ee-71db-4915-9988-cb8d38feaab2 service nova] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Received event network-vif-deleted-1ea57e46-d47a-423b-8def-3a0c88199865 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 802.238038] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "a2e70590-17ed-4804-b232-57526e87d22b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.240167] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.240524] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "a2e70590-17ed-4804-b232-57526e87d22b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.240799] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.241231] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.243444] env[63175]: INFO nova.compute.manager [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Terminating instance [ 802.320130] env[63175]: INFO nova.compute.manager [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Took 0.56 seconds to detach 1 volumes for instance. [ 802.323006] env[63175]: DEBUG nova.compute.manager [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Deleting volume: ac967057-2445-49c8-8181-4d0115179902 {{(pid=63175) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 802.351097] env[63175]: DEBUG oslo_vmware.api [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248035, 'name': PowerOnVM_Task, 'duration_secs': 0.561905} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.352483] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.354490] env[63175]: INFO nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Took 9.11 seconds to spawn the instance on the hypervisor. [ 802.354761] env[63175]: DEBUG nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 802.355608] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e450a3-3ff8-4a73-bd4c-d331ec8557f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.383869] env[63175]: DEBUG nova.scheduler.client.report [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 802.462055] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 802.493408] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.493408] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.494092] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.494517] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.494841] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.495147] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.495513] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.495829] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.496182] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.497348] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.497348] env[63175]: DEBUG nova.virt.hardware [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.498966] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8ce240-2ae1-4001-ac49-f4c3302895a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.508512] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8638da69-7926-431f-b5d8-5f388de6f7e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.603262] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527d425f-0b96-ce4b-44bb-1c51fbcbeecc, 'name': SearchDatastore_Task, 'duration_secs': 0.012626} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.605034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.605034] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.605034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.605034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.605206] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.605206] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-349f2c42-5cd1-4f5c-a188-ef70528997cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.619061] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.619061] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.619061] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-230d0e49-22f1-4c3e-8cfa-dbc91e2ce6ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.629026] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 802.629026] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a4781d-8a1e-e312-6120-5fd71199a99f" [ 802.629026] env[63175]: _type = "Task" [ 802.629026] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.632579] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248037, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074925} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.636479] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.637678] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dbdf53-34fd-4493-8534-281e28065534 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.648203] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a4781d-8a1e-e312-6120-5fd71199a99f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.666471] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 72d131a0-2617-49a1-8aff-897908929bb0/72d131a0-2617-49a1-8aff-897908929bb0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.666893] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-570268a8-6eef-478f-8936-1965405dc22c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.692045] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 802.692045] env[63175]: value = "task-1248039" [ 802.692045] env[63175]: _type = "Task" [ 802.692045] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.702548] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248039, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.748191] env[63175]: DEBUG nova.compute.manager [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 802.748407] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.752021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faef5a1f-4d9c-4e99-8056-228d897c39ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.759343] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 802.759595] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3a1a165-ae83-4a7c-b2ef-938634e8978a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.766870] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 802.766870] env[63175]: value = "task-1248040" [ 802.766870] env[63175]: _type = "Task" [ 802.766870] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.775300] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248040, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.877474] env[63175]: INFO nova.compute.manager [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Took 38.96 seconds to build instance. [ 802.881467] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.887859] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.888046] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 802.890874] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.513s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.891111] env[63175]: DEBUG nova.objects.instance [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'resources' on Instance uuid 0f8e580e-fb14-4db8-b995-a9ffe06d8bac {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 803.142739] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a4781d-8a1e-e312-6120-5fd71199a99f, 'name': SearchDatastore_Task, 'duration_secs': 0.038771} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.143487] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b8c4cc5-8b16-4d54-b8f7-43e0f206cc1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.149312] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 803.149312] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfbbb3-2ef1-7b86-6c12-a4de147fa95a" [ 803.149312] env[63175]: _type = "Task" [ 803.149312] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.159507] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfbbb3-2ef1-7b86-6c12-a4de147fa95a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.201995] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.276504] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248040, 'name': PowerOffVM_Task, 'duration_secs': 0.446051} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.276793] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 803.276902] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.277288] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d39e9835-2969-4822-8f51-4bbe1aa85a4e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.299704] env[63175]: INFO nova.compute.manager [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Rebuilding instance [ 803.342078] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.342360] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.342546] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Deleting the datastore file [datastore1] a2e70590-17ed-4804-b232-57526e87d22b {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.342809] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bce096a4-b628-4be9-8e72-c564ff7d3218 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.351795] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for the task: (returnval){ [ 803.351795] env[63175]: value = "task-1248042" [ 803.351795] env[63175]: _type = "Task" [ 803.351795] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.360996] env[63175]: DEBUG nova.compute.manager [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 803.362042] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575d1f13-4383-42c9-8c1a-14ae9e04e5fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.370797] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.379355] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b8c19291-f9af-46b2-bfd0-545c0d64d089 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.233s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.393997] env[63175]: DEBUG nova.compute.utils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.395517] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 803.395716] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.443199] env[63175]: DEBUG nova.policy [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e3ed7f44704480380c91e92d7a94c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe51efa241a40a5ae4c306b1688ccae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.666709] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfbbb3-2ef1-7b86-6c12-a4de147fa95a, 'name': SearchDatastore_Task, 'duration_secs': 0.023167} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.667056] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.667360] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 4d07a537-7df6-4659-8760-bf7e7925da25/4d07a537-7df6-4659-8760-bf7e7925da25.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.669023] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89797f95-2db2-45c0-913e-1948bde19dc7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.676876] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 803.676876] env[63175]: value = "task-1248043" [ 803.676876] env[63175]: _type = "Task" [ 803.676876] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.690989] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248043, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.705733] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248039, 'name': ReconfigVM_Task, 'duration_secs': 0.707666} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.707154] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 72d131a0-2617-49a1-8aff-897908929bb0/72d131a0-2617-49a1-8aff-897908929bb0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.708208] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e7a297d-bc86-4873-9327-6d69f9026111 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.714361] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 803.714361] env[63175]: value = "task-1248044" [ 803.714361] env[63175]: _type = "Task" [ 803.714361] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.724822] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248044, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.778125] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89237ec-e11f-4dae-835e-c2ae20f64d51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.785900] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756f7028-493b-4a05-8265-7d2ccf4a6357 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.823040] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Successfully created port: 1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.825826] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b48346-495b-4744-baad-c9a45c1228dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.836449] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb1e10b-5378-4442-a54a-7a235dce93b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.854831] env[63175]: DEBUG nova.compute.provider_tree [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.867033] env[63175]: DEBUG oslo_vmware.api [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Task: {'id': task-1248042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.42297} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.867033] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.867033] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 803.867033] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.867033] env[63175]: INFO nova.compute.manager [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 803.867268] env[63175]: DEBUG oslo.service.loopingcall [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.867268] env[63175]: DEBUG nova.compute.manager [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 803.867268] env[63175]: DEBUG nova.network.neutron [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 803.899480] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 803.978397] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Successfully updated port: 57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.190179] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248043, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.227842] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248044, 'name': Rename_Task, 'duration_secs': 0.157529} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.228607] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.228607] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb140f13-32d9-4363-8185-2d1d9ed02a4d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.235069] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 804.235069] env[63175]: value = "task-1248045" [ 804.235069] env[63175]: _type = "Task" [ 804.235069] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.241328] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Received event network-changed-71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 804.241499] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Refreshing instance network info cache due to event network-changed-71756e11-f67d-4268-a4b0-25b0a8cdefbf. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 804.241774] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.241873] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.242315] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Refreshing network info cache for port 71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.249162] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.361591] env[63175]: DEBUG nova.scheduler.client.report [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 804.382277] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.383257] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef085efd-a294-482d-af75-90dc09cab753 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.390175] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 804.390175] env[63175]: value = "task-1248046" [ 804.390175] env[63175]: _type = "Task" [ 804.390175] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.401613] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248046, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.481157] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.481380] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.481547] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.688468] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248043, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550011} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.688742] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 4d07a537-7df6-4659-8760-bf7e7925da25/4d07a537-7df6-4659-8760-bf7e7925da25.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.688959] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.689226] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-652a091e-f077-40d5-85b1-90d5b226ce06 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.696741] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 804.696741] env[63175]: value = "task-1248047" [ 804.696741] env[63175]: _type = "Task" [ 804.696741] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.704107] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248047, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.745291] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248045, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.819325] env[63175]: DEBUG nova.network.neutron [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.867054] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.869705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.912s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.870208] env[63175]: DEBUG nova.objects.instance [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lazy-loading 'resources' on Instance uuid 67e8715d-b729-4013-8cca-44eaa55c662e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.896144] env[63175]: INFO nova.scheduler.client.report [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance 0f8e580e-fb14-4db8-b995-a9ffe06d8bac [ 804.903927] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248046, 'name': PowerOffVM_Task, 'duration_secs': 0.221337} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.904295] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.904593] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.905407] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c529ecc-5175-46bd-9fae-772a7d6e9739 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.908916] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 804.915652] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.916028] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d27dcd7-bff2-4999-a3f5-464798e00999 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.946384] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.947095] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.947654] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.947942] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.948189] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.948409] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.948762] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.948993] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.949284] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.949523] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.949743] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.951141] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903e8e82-79e4-413e-97c7-071ea01a12c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.965240] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b32c0c6-e590-4d47-a23a-5efd5348d0a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.064722] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.082821] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.083124] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.083351] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleting the datastore file [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.083660] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8db8ced-fb7e-4835-a024-e218a4c64682 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.090255] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 805.090255] env[63175]: value = "task-1248049" [ 805.090255] env[63175]: _type = "Task" [ 805.090255] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.098758] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248049, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.209810] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248047, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067814} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.212255] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.213129] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1607ac2b-4ab3-471f-ad8d-53e2f55d67c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.241611] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 4d07a537-7df6-4659-8760-bf7e7925da25/4d07a537-7df6-4659-8760-bf7e7925da25.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.250602] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a49f110-5417-43e4-8f52-573dcadb1973 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.270109] env[63175]: DEBUG oslo_vmware.api [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248045, 'name': PowerOnVM_Task, 'duration_secs': 1.031417} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.271384] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.271604] env[63175]: INFO nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Took 9.66 seconds to spawn the instance on the hypervisor. [ 805.271786] env[63175]: DEBUG nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 805.272142] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 805.272142] env[63175]: value = "task-1248050" [ 805.272142] env[63175]: _type = "Task" [ 805.272142] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.272813] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737d66a5-3202-40d2-9c3c-ada8de7b912f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.286530] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.322896] env[63175]: INFO nova.compute.manager [-] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Took 1.46 seconds to deallocate network for instance. [ 805.408011] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e9ef50ad-f6d8-48f8-8b29-1e13119afb26 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0f8e580e-fb14-4db8-b995-a9ffe06d8bac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.853s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.452889] env[63175]: DEBUG nova.compute.manager [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Received event network-vif-plugged-1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 805.453188] env[63175]: DEBUG oslo_concurrency.lockutils [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] Acquiring lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.455147] env[63175]: DEBUG oslo_concurrency.lockutils [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] Lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.455147] env[63175]: DEBUG oslo_concurrency.lockutils [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] Lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.455147] env[63175]: DEBUG nova.compute.manager [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] No waiting events found dispatching network-vif-plugged-1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 805.455147] env[63175]: WARNING nova.compute.manager [req-514749a3-d388-4599-83be-34b884680f42 req-c51107d0-9eb8-402d-a87e-29b11d1c9993 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Received unexpected event network-vif-plugged-1a07e685-ea5f-4104-8423-b0a5b0aa6a82 for instance with vm_state building and task_state spawning. [ 805.488878] env[63175]: DEBUG nova.network.neutron [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating instance_info_cache with network_info: [{"id": "57912c9e-afb5-4a48-af4e-c569456da08b", "address": "fa:16:3e:44:b0:e4", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57912c9e-af", "ovs_interfaceid": "57912c9e-afb5-4a48-af4e-c569456da08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.527825] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Successfully updated port: 1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.570013] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updated VIF entry in instance network info cache for port 71756e11-f67d-4268-a4b0-25b0a8cdefbf. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.570013] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.603898] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248049, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.662592] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1265b3-fb40-4755-9997-97d88e2a2f60 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.670205] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca46264-a177-4898-b42e-6100e1cf6e4a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.700729] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f4dc1e-0a8d-41d9-9b3b-088aec0f074c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.708160] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4034e9-1cd4-428d-a498-4ccd9066690e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.721813] env[63175]: DEBUG nova.compute.provider_tree [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.784663] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.799123] env[63175]: INFO nova.compute.manager [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Took 39.80 seconds to build instance. [ 805.832955] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.991802] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.992391] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Instance network_info: |[{"id": "57912c9e-afb5-4a48-af4e-c569456da08b", "address": "fa:16:3e:44:b0:e4", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57912c9e-af", "ovs_interfaceid": "57912c9e-afb5-4a48-af4e-c569456da08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 805.992636] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:b0:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57912c9e-afb5-4a48-af4e-c569456da08b', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.000081] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating folder: Project (54aee75c86d543a995a364ed78426ec2). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.000380] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f5a1e72-3a22-485a-b53e-fff39309548a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.011792] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created folder: Project (54aee75c86d543a995a364ed78426ec2) in parent group-v268956. [ 806.012045] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating folder: Instances. Parent ref: group-v269014. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.012314] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b701d84c-3e29-4e7c-897a-d6e24b24f76f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.021140] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created folder: Instances in parent group-v269014. [ 806.021512] env[63175]: DEBUG oslo.service.loopingcall [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.021723] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.021933] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e937a3d-dfe3-405c-b018-518035e3d7fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.036772] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.036921] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.037148] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.043406] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.043406] env[63175]: value = "task-1248053" [ 806.043406] env[63175]: _type = "Task" [ 806.043406] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.059967] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248053, 'name': CreateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.071128] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.071461] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Received event network-vif-plugged-57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 806.071730] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.072031] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.072224] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.072391] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] No waiting events found dispatching network-vif-plugged-57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 806.072551] env[63175]: WARNING nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Received unexpected event network-vif-plugged-57912c9e-afb5-4a48-af4e-c569456da08b for instance with vm_state building and task_state spawning. [ 806.072706] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Received event network-changed-57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 806.072852] env[63175]: DEBUG nova.compute.manager [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Refreshing instance network info cache due to event network-changed-57912c9e-afb5-4a48-af4e-c569456da08b. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 806.073055] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Acquiring lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.073194] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Acquired lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.073344] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Refreshing network info cache for port 57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.103853] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248049, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.640343} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.104353] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.104640] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.104978] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.225344] env[63175]: DEBUG nova.scheduler.client.report [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 806.288313] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248050, 'name': ReconfigVM_Task, 'duration_secs': 0.785688} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.289188] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 4d07a537-7df6-4659-8760-bf7e7925da25/4d07a537-7df6-4659-8760-bf7e7925da25.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.289733] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5c2d84d-e4ac-493e-b2ba-b6495f00e854 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.297529] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 806.297529] env[63175]: value = "task-1248054" [ 806.297529] env[63175]: _type = "Task" [ 806.297529] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.302094] env[63175]: DEBUG oslo_concurrency.lockutils [None req-28ddc297-e149-4489-ab1f-bc15fb8423bd tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.350s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.307734] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248054, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.371106] env[63175]: DEBUG nova.compute.manager [req-25c375a2-8c00-4036-bd93-bea4c01b01a9 req-0b3fe289-377a-449d-9250-da65e168ab18 service nova] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Received event network-vif-deleted-5301b2d9-86e8-4b5b-abd3-6443a9d5be1a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 806.556367] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248053, 'name': CreateVM_Task, 'duration_secs': 0.34958} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.556367] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.556367] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.556367] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.556367] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.556735] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4800d169-962e-4e0f-b2ea-f0d38251f7b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.561411] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 806.561411] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d8697d-8003-311c-ea3f-97a9e53b1b15" [ 806.561411] env[63175]: _type = "Task" [ 806.561411] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.573339] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d8697d-8003-311c-ea3f-97a9e53b1b15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.573339] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.732788] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.737046] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.860s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.737246] env[63175]: DEBUG nova.objects.instance [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 806.754912] env[63175]: INFO nova.scheduler.client.report [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Deleted allocations for instance 67e8715d-b729-4013-8cca-44eaa55c662e [ 806.795991] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Updating instance_info_cache with network_info: [{"id": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "address": "fa:16:3e:d0:5a:21", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a07e685-ea", "ovs_interfaceid": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.809967] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248054, 'name': Rename_Task, 'duration_secs': 0.257717} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.811163] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.812339] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6444706-df7b-4cc4-bc16-02241c53aafa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.819596] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 806.819596] env[63175]: value = "task-1248055" [ 806.819596] env[63175]: _type = "Task" [ 806.819596] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.832277] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.978545] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updated VIF entry in instance network info cache for port 57912c9e-afb5-4a48-af4e-c569456da08b. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.978545] env[63175]: DEBUG nova.network.neutron [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating instance_info_cache with network_info: [{"id": "57912c9e-afb5-4a48-af4e-c569456da08b", "address": "fa:16:3e:44:b0:e4", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57912c9e-af", "ovs_interfaceid": "57912c9e-afb5-4a48-af4e-c569456da08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.076040] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d8697d-8003-311c-ea3f-97a9e53b1b15, 'name': SearchDatastore_Task, 'duration_secs': 0.009801} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.076040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.076040] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.076040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.076446] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.076446] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.076446] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f289d78-8226-4c5e-9871-53ba5c4f513d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.087941] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.087941] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.087941] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae09fe7d-ac17-49e7-8aae-f00ef8b1f663 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.092738] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 807.092738] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ac98d1-0a05-d295-1eee-9965ad8d214e" [ 807.092738] env[63175]: _type = "Task" [ 807.092738] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.107927] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ac98d1-0a05-d295-1eee-9965ad8d214e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.114593] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "72d131a0-2617-49a1-8aff-897908929bb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.115416] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.115416] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "72d131a0-2617-49a1-8aff-897908929bb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.115416] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.115837] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.119377] env[63175]: INFO nova.compute.manager [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Terminating instance [ 807.157666] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.158101] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.158357] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.158715] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.159042] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.159437] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.159926] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.160276] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.160632] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.160981] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.161368] env[63175]: DEBUG nova.virt.hardware [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.163230] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9b75e6-3752-4aac-886b-9db353ba64eb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.172946] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1db8650-5f17-4ea2-89bd-e13ce07b7334 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.198245] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:9c:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.210243] env[63175]: DEBUG oslo.service.loopingcall [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.214017] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.214017] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-765b688f-9fb4-4d3f-ac2e-072576565a08 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.233654] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.233654] env[63175]: value = "task-1248056" [ 807.233654] env[63175]: _type = "Task" [ 807.233654] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.245743] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248056, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.265039] env[63175]: DEBUG oslo_concurrency.lockutils [None req-264c1377-17b8-4473-8796-4dbd831bb0f2 tempest-FloatingIPsAssociationNegativeTestJSON-319913746 tempest-FloatingIPsAssociationNegativeTestJSON-319913746-project-member] Lock "67e8715d-b729-4013-8cca-44eaa55c662e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.665s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.297298] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.297716] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Instance network_info: |[{"id": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "address": "fa:16:3e:d0:5a:21", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a07e685-ea", "ovs_interfaceid": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 807.298243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:5a:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '23f4655e-3495-421d-be4e-f6002a85a47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a07e685-ea5f-4104-8423-b0a5b0aa6a82', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.307995] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Creating folder: Project (afe51efa241a40a5ae4c306b1688ccae). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.308714] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10ebeb36-e819-4374-9763-43f5b7ff86ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.319837] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Created folder: Project (afe51efa241a40a5ae4c306b1688ccae) in parent group-v268956. [ 807.320094] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Creating folder: Instances. Parent ref: group-v269018. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.320380] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af596923-ddd0-434f-9709-03cfbcca5154 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.331870] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248055, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.333581] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Created folder: Instances in parent group-v269018. [ 807.333838] env[63175]: DEBUG oslo.service.loopingcall [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.334047] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.334509] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5bb380d-35f4-4615-9a63-8f5f03ee9870 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.355521] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.355521] env[63175]: value = "task-1248059" [ 807.355521] env[63175]: _type = "Task" [ 807.355521] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.368158] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248059, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.483878] env[63175]: DEBUG oslo_concurrency.lockutils [req-fcb10ef3-80ba-4a81-b43d-aa9a4b195f31 req-0afb387d-f2ef-4de7-8c66-e8fcc29e0a1a service nova] Releasing lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.589833] env[63175]: DEBUG nova.compute.manager [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Received event network-changed-1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 807.590108] env[63175]: DEBUG nova.compute.manager [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Refreshing instance network info cache due to event network-changed-1a07e685-ea5f-4104-8423-b0a5b0aa6a82. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 807.591964] env[63175]: DEBUG oslo_concurrency.lockutils [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] Acquiring lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.591964] env[63175]: DEBUG oslo_concurrency.lockutils [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] Acquired lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.592154] env[63175]: DEBUG nova.network.neutron [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Refreshing network info cache for port 1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.605245] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ac98d1-0a05-d295-1eee-9965ad8d214e, 'name': SearchDatastore_Task, 'duration_secs': 0.009063} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.606606] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f053fc9-e77f-4e5d-af81-de656447b030 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.613160] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 807.613160] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52587bdc-62d6-4777-65ea-71b19ac34b2d" [ 807.613160] env[63175]: _type = "Task" [ 807.613160] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.629296] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52587bdc-62d6-4777-65ea-71b19ac34b2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.629780] env[63175]: DEBUG nova.compute.manager [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.629989] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.631121] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4eb3c29-4f6b-43b9-9a6c-188b937f33ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.644309] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.644309] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47fb1fa3-82bb-465f-b115-73cf19a2c89b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.649092] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 807.649092] env[63175]: value = "task-1248060" [ 807.649092] env[63175]: _type = "Task" [ 807.649092] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.658718] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.744666] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248056, 'name': CreateVM_Task, 'duration_secs': 0.400452} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.744838] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.745558] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.745714] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.746055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.747186] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25febe75-b65c-4c24-bcfc-641eed5a1c68 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.748147] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2157a15-cabe-4bc6-af04-26e4efe47cc8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.750186] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.684s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.750409] env[63175]: DEBUG nova.objects.instance [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lazy-loading 'resources' on Instance uuid 36089589-d105-49e0-8ae7-790c814b036c {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.757717] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 807.757717] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520202d3-bddd-3ebb-a1f5-e7c115adf688" [ 807.757717] env[63175]: _type = "Task" [ 807.757717] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.770518] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520202d3-bddd-3ebb-a1f5-e7c115adf688, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.832619] env[63175]: DEBUG oslo_vmware.api [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248055, 'name': PowerOnVM_Task, 'duration_secs': 0.718868} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.833173] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.833534] env[63175]: INFO nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Took 9.66 seconds to spawn the instance on the hypervisor. [ 807.833856] env[63175]: DEBUG nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 807.837021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff81572d-24ba-4a89-beb1-ef3a081953e1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.868234] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248059, 'name': CreateVM_Task, 'duration_secs': 0.347679} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.868234] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.868234] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.868234] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.868234] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.868598] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61f9bf38-6d2d-4941-b1d1-76fbee9a0e26 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.875719] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 807.875719] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52870b45-d758-354d-0f79-9ea0cf258ee3" [ 807.875719] env[63175]: _type = "Task" [ 807.875719] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.891309] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52870b45-d758-354d-0f79-9ea0cf258ee3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.123856] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52587bdc-62d6-4777-65ea-71b19ac34b2d, 'name': SearchDatastore_Task, 'duration_secs': 0.010285} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.124147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.124409] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] e1e4f169-000c-4e9c-8ef5-aa4b4989eb44/e1e4f169-000c-4e9c-8ef5-aa4b4989eb44.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.124672] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-415bf684-0885-4c04-8648-abf84f776ebf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.137042] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 808.137042] env[63175]: value = "task-1248061" [ 808.137042] env[63175]: _type = "Task" [ 808.137042] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.145575] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248061, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.161020] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248060, 'name': PowerOffVM_Task, 'duration_secs': 0.254639} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.161020] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.161020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.161020] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-114b9f41-f46a-49f4-aee0-d77bfef3f8de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.245439] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.245439] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.245439] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Deleting the datastore file [datastore2] 72d131a0-2617-49a1-8aff-897908929bb0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.245439] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c7ceb8c-a454-4ec6-b830-482e72ef426a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.253030] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for the task: (returnval){ [ 808.253030] env[63175]: value = "task-1248063" [ 808.253030] env[63175]: _type = "Task" [ 808.253030] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.272230] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.283960] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520202d3-bddd-3ebb-a1f5-e7c115adf688, 'name': SearchDatastore_Task, 'duration_secs': 0.013297} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.284299] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.284536] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.284765] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.284919] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.285104] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.285379] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c806496c-d814-4e43-bf46-d6a7a922c793 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.302385] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.302567] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.303378] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a470f9e5-f4cb-4f1a-9f82-040823dc8d19 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.308642] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 808.308642] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c75bac-33c8-e014-7205-90828d4405bb" [ 808.308642] env[63175]: _type = "Task" [ 808.308642] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.318933] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c75bac-33c8-e014-7205-90828d4405bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.356378] env[63175]: INFO nova.compute.manager [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Took 39.26 seconds to build instance. [ 808.386778] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52870b45-d758-354d-0f79-9ea0cf258ee3, 'name': SearchDatastore_Task, 'duration_secs': 0.03104} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.387201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.387315] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.387544] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.387688] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.387862] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.393230] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0c0d688-f104-4493-8243-eb8dec246575 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.405855] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.406074] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.412255] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06179367-1654-4ad7-85c2-ee64b5d25005 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.422618] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 808.422618] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52697440-8448-797a-082d-df911395019d" [ 808.422618] env[63175]: _type = "Task" [ 808.422618] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.435539] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52697440-8448-797a-082d-df911395019d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.443906] env[63175]: DEBUG nova.network.neutron [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Updated VIF entry in instance network info cache for port 1a07e685-ea5f-4104-8423-b0a5b0aa6a82. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 808.443906] env[63175]: DEBUG nova.network.neutron [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Updating instance_info_cache with network_info: [{"id": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "address": "fa:16:3e:d0:5a:21", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a07e685-ea", "ovs_interfaceid": "1a07e685-ea5f-4104-8423-b0a5b0aa6a82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.609267] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.609583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.618035] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce6a852-43e6-45cd-afe9-0d348544b8f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.630732] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e459b6d-ab35-4713-b235-c4d477d31157 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.670421] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa57ad5f-e6d3-4d39-8628-ddd6bd924aed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.683128] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248061, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.684400] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0911d572-04e6-4ea5-9526-060d5962772a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.699131] env[63175]: DEBUG nova.compute.provider_tree [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.762458] env[63175]: DEBUG oslo_vmware.api [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Task: {'id': task-1248063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.410622} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.762719] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.762955] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.763209] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.763465] env[63175]: INFO nova.compute.manager [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 808.763765] env[63175]: DEBUG oslo.service.loopingcall [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.763980] env[63175]: DEBUG nova.compute.manager [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 808.764094] env[63175]: DEBUG nova.network.neutron [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.819357] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c75bac-33c8-e014-7205-90828d4405bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009786} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.820212] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57cb5db8-a560-49f9-8861-569a7cc26c58 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.826921] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 808.826921] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52665d0c-7d65-38ba-64d5-dd7bef46acb7" [ 808.826921] env[63175]: _type = "Task" [ 808.826921] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.837625] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52665d0c-7d65-38ba-64d5-dd7bef46acb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.858474] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cc85907a-e58b-4606-9d43-39af07f355b3 tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.933786] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52697440-8448-797a-082d-df911395019d, 'name': SearchDatastore_Task, 'duration_secs': 0.014123} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.934665] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2787a9c-d60a-4370-819f-dfe4aa68e3a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.940213] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 808.940213] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfd17d-586c-725a-029e-43b874e9e431" [ 808.940213] env[63175]: _type = "Task" [ 808.940213] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.947892] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfd17d-586c-725a-029e-43b874e9e431, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.950686] env[63175]: DEBUG oslo_concurrency.lockutils [req-265a1007-0d00-4870-8599-62840a50ed7f req-52237e20-d937-43c2-8f19-3ba1e69dabb8 service nova] Releasing lock "refresh_cache-780989ee-98a1-4e99-9014-45de6921d4bc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.113275] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 809.176625] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248061, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705255} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.176959] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] e1e4f169-000c-4e9c-8ef5-aa4b4989eb44/e1e4f169-000c-4e9c-8ef5-aa4b4989eb44.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.177241] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.177612] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68d67090-9324-4d6d-a28e-cedb6fb9d695 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.186230] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 809.186230] env[63175]: value = "task-1248064" [ 809.186230] env[63175]: _type = "Task" [ 809.186230] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.194452] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.207265] env[63175]: DEBUG nova.scheduler.client.report [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 809.337197] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52665d0c-7d65-38ba-64d5-dd7bef46acb7, 'name': SearchDatastore_Task, 'duration_secs': 0.015415} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.338684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.338684] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.338875] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c982875-c7f9-44b2-a3f5-3bfcb8657659 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.345964] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 809.345964] env[63175]: value = "task-1248065" [ 809.345964] env[63175]: _type = "Task" [ 809.345964] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.355846] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.452498] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bfd17d-586c-725a-029e-43b874e9e431, 'name': SearchDatastore_Task, 'duration_secs': 0.026854} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.452498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.452498] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 780989ee-98a1-4e99-9014-45de6921d4bc/780989ee-98a1-4e99-9014-45de6921d4bc.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.452498] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fc9dc31-a010-4e3e-b5fe-822178470ede {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.458154] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 809.458154] env[63175]: value = "task-1248066" [ 809.458154] env[63175]: _type = "Task" [ 809.458154] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.466030] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.641057] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.669723] env[63175]: DEBUG nova.compute.manager [req-244b88f4-ab56-4d99-bf39-532207f98a75 req-4883d371-0bc5-4fbd-af31-a6001fca7931 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Received event network-vif-deleted-810e244a-4573-4ae6-98d1-debc881e4ab2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 809.669934] env[63175]: INFO nova.compute.manager [req-244b88f4-ab56-4d99-bf39-532207f98a75 req-4883d371-0bc5-4fbd-af31-a6001fca7931 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Neutron deleted interface 810e244a-4573-4ae6-98d1-debc881e4ab2; detaching it from the instance and deleting it from the info cache [ 809.670154] env[63175]: DEBUG nova.network.neutron [req-244b88f4-ab56-4d99-bf39-532207f98a75 req-4883d371-0bc5-4fbd-af31-a6001fca7931 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.696429] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070847} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.697449] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.698140] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11eade3-9bec-4390-809c-debf8dcd7c21 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.717732] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.967s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.728258] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] e1e4f169-000c-4e9c-8ef5-aa4b4989eb44/e1e4f169-000c-4e9c-8ef5-aa4b4989eb44.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.728882] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.560s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.730437] env[63175]: INFO nova.compute.claims [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.733238] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d53c4e3a-6d60-46d1-b6a3-dc314f342c03 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.751209] env[63175]: DEBUG nova.network.neutron [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.758217] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 809.758217] env[63175]: value = "task-1248067" [ 809.758217] env[63175]: _type = "Task" [ 809.758217] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.765385] env[63175]: INFO nova.scheduler.client.report [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Deleted allocations for instance 36089589-d105-49e0-8ae7-790c814b036c [ 809.773847] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248067, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.858131] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248065, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.969492] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248066, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.173321] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f44c076-714c-4ea0-b1ba-28e620b4376c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.182873] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7460a7d4-96bf-4190-aff1-8ed6b1760dda {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.218022] env[63175]: DEBUG nova.compute.manager [req-244b88f4-ab56-4d99-bf39-532207f98a75 req-4883d371-0bc5-4fbd-af31-a6001fca7931 service nova] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Detach interface failed, port_id=810e244a-4573-4ae6-98d1-debc881e4ab2, reason: Instance 72d131a0-2617-49a1-8aff-897908929bb0 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 810.257124] env[63175]: INFO nova.compute.manager [-] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Took 1.49 seconds to deallocate network for instance. [ 810.277272] env[63175]: DEBUG oslo_concurrency.lockutils [None req-92974a18-3477-432f-bd01-1537a4b55444 tempest-ServerShowV254Test-639799367 tempest-ServerShowV254Test-639799367-project-member] Lock "36089589-d105-49e0-8ae7-790c814b036c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.439s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.278596] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248067, 'name': ReconfigVM_Task, 'duration_secs': 0.355833} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.278891] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfigured VM instance instance-0000003d to attach disk [datastore2] e1e4f169-000c-4e9c-8ef5-aa4b4989eb44/e1e4f169-000c-4e9c-8ef5-aa4b4989eb44.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.279858] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1750862f-f215-4d35-8a5b-d9b81e58b346 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.288397] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 810.288397] env[63175]: value = "task-1248068" [ 810.288397] env[63175]: _type = "Task" [ 810.288397] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.300843] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248068, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.364457] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248065, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.962472} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.364806] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 810.365073] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.365903] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e317ed5-3253-4d0b-9aac-35b65c3eca43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.372164] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 810.372164] env[63175]: value = "task-1248069" [ 810.372164] env[63175]: _type = "Task" [ 810.372164] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.385278] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248069, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.469602] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248066, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.768527] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.802773] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248068, 'name': Rename_Task, 'duration_secs': 0.21506} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.803077] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.803344] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67a715a0-323e-415f-8260-865f33b5b0b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.810262] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 810.810262] env[63175]: value = "task-1248070" [ 810.810262] env[63175]: _type = "Task" [ 810.810262] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.819015] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248070, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.881881] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248069, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061498} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.885873] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.887062] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76ea679-2147-462b-be74-ada638cbec8e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.918315] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.918654] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2835b815-7814-4d09-a4af-b6b78c824b5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.941512] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 810.941512] env[63175]: value = "task-1248071" [ 810.941512] env[63175]: _type = "Task" [ 810.941512] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.954273] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248071, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.970307] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248066, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.048030] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6cdba2-4dde-43c9-8cf5-436be2724597 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.055577] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f98199-6f85-43bc-9043-3010f850f68c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.085261] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9778fa65-775f-4681-9d71-2429f6497fb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.092752] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26f45e0-7759-4016-b390-30a639e05dfc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.106102] env[63175]: DEBUG nova.compute.provider_tree [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.321125] env[63175]: DEBUG oslo_vmware.api [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248070, 'name': PowerOnVM_Task, 'duration_secs': 0.447683} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.321479] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.321638] env[63175]: INFO nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Took 8.86 seconds to spawn the instance on the hypervisor. [ 811.321787] env[63175]: DEBUG nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.322625] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ef97e1-f0fb-47ab-8020-14a64fac3b52 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.451321] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248071, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.468866] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248066, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.914107} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.469011] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 780989ee-98a1-4e99-9014-45de6921d4bc/780989ee-98a1-4e99-9014-45de6921d4bc.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.469212] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.469423] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c7b697a-711a-448c-a1cb-bc0a773cd3f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.475995] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 811.475995] env[63175]: value = "task-1248072" [ 811.475995] env[63175]: _type = "Task" [ 811.475995] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.483385] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.609460] env[63175]: DEBUG nova.scheduler.client.report [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 811.749149] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "4d07a537-7df6-4659-8760-bf7e7925da25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.749476] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.749809] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.751914] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.751914] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.759571] env[63175]: INFO nova.compute.manager [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Terminating instance [ 811.846054] env[63175]: INFO nova.compute.manager [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Took 36.11 seconds to build instance. [ 811.952423] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248071, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.984926] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.338991} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.985214] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.985995] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d99213-7c01-4490-9d3b-a4c7277122df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.007770] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 780989ee-98a1-4e99-9014-45de6921d4bc/780989ee-98a1-4e99-9014-45de6921d4bc.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.008010] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ba778a3-67a2-4e01-a5e5-4373e040e0c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.026875] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 812.026875] env[63175]: value = "task-1248073" [ 812.026875] env[63175]: _type = "Task" [ 812.026875] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.034795] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248073, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.114918] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.115459] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.117972] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.618s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.119486] env[63175]: INFO nova.compute.claims [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.263789] env[63175]: DEBUG nova.compute.manager [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 812.264086] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.265920] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7c4f93-06f0-42f3-bd35-887f9b88deac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.273237] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.273706] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5330953-7f3b-445a-bb2c-95882d15e8db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.281197] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 812.281197] env[63175]: value = "task-1248074" [ 812.281197] env[63175]: _type = "Task" [ 812.281197] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.291112] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248074, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.345995] env[63175]: DEBUG nova.compute.manager [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Received event network-changed-57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 812.346861] env[63175]: DEBUG nova.compute.manager [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Refreshing instance network info cache due to event network-changed-57912c9e-afb5-4a48-af4e-c569456da08b. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 812.346861] env[63175]: DEBUG oslo_concurrency.lockutils [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] Acquiring lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.346861] env[63175]: DEBUG oslo_concurrency.lockutils [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] Acquired lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.346861] env[63175]: DEBUG nova.network.neutron [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Refreshing network info cache for port 57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.348603] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c4ee8bd7-0484-4250-89cb-cd5a3af3cc24 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.962s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.453110] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248071, 'name': ReconfigVM_Task, 'duration_secs': 1.291883} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.453110] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Reconfigured VM instance instance-00000039 to attach disk [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0/f6d2c297-1cfe-4cab-8854-eb5760e6cbc0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.453580] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30641a74-8518-44b3-8d59-750d9c1c9c44 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.461634] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 812.461634] env[63175]: value = "task-1248075" [ 812.461634] env[63175]: _type = "Task" [ 812.461634] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.473903] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248075, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.542028] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248073, 'name': ReconfigVM_Task, 'duration_secs': 0.337351} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.542659] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 780989ee-98a1-4e99-9014-45de6921d4bc/780989ee-98a1-4e99-9014-45de6921d4bc.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.543568] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45cbde3f-3cca-4069-b99a-8430184353ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.549384] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 812.549384] env[63175]: value = "task-1248076" [ 812.549384] env[63175]: _type = "Task" [ 812.549384] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.558019] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248076, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.625710] env[63175]: DEBUG nova.compute.utils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.629524] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 812.629762] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.674872] env[63175]: DEBUG nova.policy [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e3ed7f44704480380c91e92d7a94c81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe51efa241a40a5ae4c306b1688ccae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.792257] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248074, 'name': PowerOffVM_Task, 'duration_secs': 0.210253} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.792257] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.792257] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 812.792257] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af40b92a-0781-42d7-b2f8-b00f27eab1dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.855900] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.856500] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.856806] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Deleting the datastore file [datastore2] 4d07a537-7df6-4659-8760-bf7e7925da25 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.857605] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3e1edde-3f38-4e06-a95d-653490abdba8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.867575] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for the task: (returnval){ [ 812.867575] env[63175]: value = "task-1248078" [ 812.867575] env[63175]: _type = "Task" [ 812.867575] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.880461] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.972674] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248075, 'name': Rename_Task, 'duration_secs': 0.142783} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.972950] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.973228] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e646ebc7-fa23-46dc-88f4-c03aa82986f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.979804] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 812.979804] env[63175]: value = "task-1248079" [ 812.979804] env[63175]: _type = "Task" [ 812.979804] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.987023] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.992911] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Successfully created port: 81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.059070] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248076, 'name': Rename_Task, 'duration_secs': 0.141803} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.059349] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.059594] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d291e5f-383a-42ba-b2d2-27cd8796117f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.064877] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 813.064877] env[63175]: value = "task-1248080" [ 813.064877] env[63175]: _type = "Task" [ 813.064877] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.073711] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.130335] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 813.284384] env[63175]: DEBUG nova.network.neutron [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updated VIF entry in instance network info cache for port 57912c9e-afb5-4a48-af4e-c569456da08b. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.284841] env[63175]: DEBUG nova.network.neutron [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating instance_info_cache with network_info: [{"id": "57912c9e-afb5-4a48-af4e-c569456da08b", "address": "fa:16:3e:44:b0:e4", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57912c9e-af", "ovs_interfaceid": "57912c9e-afb5-4a48-af4e-c569456da08b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.377029] env[63175]: DEBUG oslo_vmware.api [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Task: {'id': task-1248078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134733} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.377311] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.377499] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 813.377682] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 813.377853] env[63175]: INFO nova.compute.manager [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Took 1.11 seconds to destroy the instance on the hypervisor. [ 813.378148] env[63175]: DEBUG oslo.service.loopingcall [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.378356] env[63175]: DEBUG nova.compute.manager [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 813.378450] env[63175]: DEBUG nova.network.neutron [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.479965] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57684ffb-16b9-4260-b8f1-daa251558b8b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.491674] env[63175]: DEBUG oslo_vmware.api [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248079, 'name': PowerOnVM_Task, 'duration_secs': 0.449671} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.493540] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.493759] env[63175]: DEBUG nova.compute.manager [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 813.494561] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5787a935-72de-4ec7-a457-fff7737e42bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.497722] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31eb9d8f-2c23-48b2-85c6-7a60193eb3f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.534978] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a82768-a41f-485a-8bac-76a885bdcae2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.544702] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322976ae-87e0-4a71-adb6-090f4d158c88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.559153] env[63175]: DEBUG nova.compute.provider_tree [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.577382] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248080, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.710256] env[63175]: DEBUG nova.compute.manager [req-4be9cbe9-3dd6-46a5-a993-91ec30b9f244 req-1aae9cd0-af2c-47cb-bc35-a95ad2296476 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Received event network-vif-deleted-ef671ac5-130d-4e3e-a8e6-2650bbf1c278 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 813.710256] env[63175]: INFO nova.compute.manager [req-4be9cbe9-3dd6-46a5-a993-91ec30b9f244 req-1aae9cd0-af2c-47cb-bc35-a95ad2296476 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Neutron deleted interface ef671ac5-130d-4e3e-a8e6-2650bbf1c278; detaching it from the instance and deleting it from the info cache [ 813.710328] env[63175]: DEBUG nova.network.neutron [req-4be9cbe9-3dd6-46a5-a993-91ec30b9f244 req-1aae9cd0-af2c-47cb-bc35-a95ad2296476 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.789886] env[63175]: DEBUG oslo_concurrency.lockutils [req-a97ac582-6c72-49e8-9400-93977877e0e1 req-387f505e-8408-4ee0-bb3b-ccc2f0f0bf79 service nova] Releasing lock "refresh_cache-e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.042413] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.062624] env[63175]: DEBUG nova.scheduler.client.report [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 814.078381] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248080, 'name': PowerOnVM_Task, 'duration_secs': 0.789806} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.078725] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.078945] env[63175]: INFO nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Took 9.17 seconds to spawn the instance on the hypervisor. [ 814.079169] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 814.080189] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcf4dbc-6b11-4e06-9619-b120bb9744a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.104638] env[63175]: DEBUG nova.network.neutron [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.142266] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.191158] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.191158] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.191158] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.191308] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.191308] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.191308] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.191308] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.191308] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.191462] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.191554] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.191674] env[63175]: DEBUG nova.virt.hardware [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.192648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ef74d1-6a11-4d0c-8d4d-8fe4c79b7264 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.204675] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06203075-22c4-4a8e-af28-a8b5b3b84d32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.225818] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3e36e4f-7f3a-4f4f-ba47-5941b36167b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.235526] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b86405-446f-415c-b3f4-683239226f07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.264690] env[63175]: DEBUG nova.compute.manager [req-4be9cbe9-3dd6-46a5-a993-91ec30b9f244 req-1aae9cd0-af2c-47cb-bc35-a95ad2296476 service nova] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Detach interface failed, port_id=ef671ac5-130d-4e3e-a8e6-2650bbf1c278, reason: Instance 4d07a537-7df6-4659-8760-bf7e7925da25 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 814.570601] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.571153] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 814.573951] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.738s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.574193] env[63175]: DEBUG nova.objects.instance [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lazy-loading 'resources' on Instance uuid 39e55060-73a7-4dbd-96cf-bc48d8737c1c {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.605724] env[63175]: INFO nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Took 37.87 seconds to build instance. [ 814.607035] env[63175]: INFO nova.compute.manager [-] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Took 1.23 seconds to deallocate network for instance. [ 814.700563] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.701760] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.707583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.707583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.707583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.709153] env[63175]: INFO nova.compute.manager [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Terminating instance [ 814.720061] env[63175]: DEBUG nova.compute.manager [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Received event network-vif-plugged-81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 814.720061] env[63175]: DEBUG oslo_concurrency.lockutils [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] Acquiring lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.720061] env[63175]: DEBUG oslo_concurrency.lockutils [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.720061] env[63175]: DEBUG oslo_concurrency.lockutils [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.720061] env[63175]: DEBUG nova.compute.manager [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] No waiting events found dispatching network-vif-plugged-81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 814.720452] env[63175]: WARNING nova.compute.manager [req-5f5ca26b-283d-4cac-b1d4-88285828c581 req-9f490703-6493-4fbb-b8c8-392dbda6c6f5 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Received unexpected event network-vif-plugged-81942ffb-5329-4928-a222-00d5cb4e5d43 for instance with vm_state building and task_state spawning. [ 814.818649] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Successfully updated port: 81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.078123] env[63175]: DEBUG nova.compute.utils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.082502] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 815.082686] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.108176] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.626s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.112479] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.166683] env[63175]: DEBUG nova.policy [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '816341e295cb4cdab052d643ad9a44d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ea6da3858c347d887dcfba2556fabad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.217026] env[63175]: DEBUG nova.compute.manager [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 815.217026] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.217026] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be48a72b-88a6-4e68-ae8d-78f64b95a449 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.230237] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.238041] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de3088a3-38a4-45a4-b3c5-445cd48da47d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.241348] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 815.241348] env[63175]: value = "task-1248081" [ 815.241348] env[63175]: _type = "Task" [ 815.241348] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.253842] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.254125] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.260367] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.323398] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.323545] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.323806] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.374109] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e637337b-ff33-4dd5-a590-ef2fde04cde7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.381089] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106443eb-c465-42a8-a8ed-91cd63f0354f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.420999] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d74ff3-ea9b-437b-a3e2-410a74a6c6c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.429260] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca437162-4617-40c5-a4d8-435c4950dafc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.443941] env[63175]: DEBUG nova.compute.provider_tree [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.583403] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 815.680341] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Successfully created port: 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.751767] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248081, 'name': PowerOffVM_Task, 'duration_secs': 0.197252} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.752827] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.754351] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.755104] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-889f0d33-17e3-4d0f-bb7f-57987b1af799 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.756375] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 815.867499] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.867775] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.868133] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleting the datastore file [datastore1] f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.868757] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9fe02fc-82f7-4cc4-a61a-273da578e7cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.875147] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 815.875147] env[63175]: value = "task-1248083" [ 815.875147] env[63175]: _type = "Task" [ 815.875147] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.879794] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.886731] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.946606] env[63175]: DEBUG nova.scheduler.client.report [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 816.125160] env[63175]: DEBUG nova.network.neutron [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Updating instance_info_cache with network_info: [{"id": "81942ffb-5329-4928-a222-00d5cb4e5d43", "address": "fa:16:3e:79:33:fc", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81942ffb-53", "ovs_interfaceid": "81942ffb-5329-4928-a222-00d5cb4e5d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.282150] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.384968] env[63175]: DEBUG oslo_vmware.api [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.31541} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.385304] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.385515] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.385735] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.385948] env[63175]: INFO nova.compute.manager [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 816.386248] env[63175]: DEBUG oslo.service.loopingcall [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.386482] env[63175]: DEBUG nova.compute.manager [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 816.386599] env[63175]: DEBUG nova.network.neutron [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.451950] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.454784] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.755s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.458141] env[63175]: INFO nova.compute.claims [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.483643] env[63175]: INFO nova.scheduler.client.report [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Deleted allocations for instance 39e55060-73a7-4dbd-96cf-bc48d8737c1c [ 816.600248] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 816.631049] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.631049] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Instance network_info: |[{"id": "81942ffb-5329-4928-a222-00d5cb4e5d43", "address": "fa:16:3e:79:33:fc", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81942ffb-53", "ovs_interfaceid": "81942ffb-5329-4928-a222-00d5cb4e5d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 816.631202] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:33:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '23f4655e-3495-421d-be4e-f6002a85a47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81942ffb-5329-4928-a222-00d5cb4e5d43', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.639044] env[63175]: DEBUG oslo.service.loopingcall [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.641207] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.641333] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.641424] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.642056] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.642056] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.642056] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.642245] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.642287] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.642436] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.642598] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.642766] env[63175]: DEBUG nova.virt.hardware [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.643048] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.643799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c735b415-d0a6-4e38-80b6-5c2bd729f7de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.646494] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d14e1a2-67ca-4b0e-847f-3128733be2b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.668202] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53207728-7462-4fca-ac8b-16a707c80aae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.673107] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.673107] env[63175]: value = "task-1248084" [ 816.673107] env[63175]: _type = "Task" [ 816.673107] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.690440] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248084, 'name': CreateVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.756682] env[63175]: DEBUG nova.compute.manager [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Received event network-changed-81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 816.756905] env[63175]: DEBUG nova.compute.manager [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Refreshing instance network info cache due to event network-changed-81942ffb-5329-4928-a222-00d5cb4e5d43. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 816.757147] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] Acquiring lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.757313] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] Acquired lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.757459] env[63175]: DEBUG nova.network.neutron [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Refreshing network info cache for port 81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.762256] env[63175]: DEBUG nova.compute.manager [req-b316d997-a552-4f9a-ab8e-5c454633100d req-1bcac634-aae8-49e1-a582-4ceea647ef58 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Received event network-vif-deleted-595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 816.762256] env[63175]: INFO nova.compute.manager [req-b316d997-a552-4f9a-ab8e-5c454633100d req-1bcac634-aae8-49e1-a582-4ceea647ef58 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Neutron deleted interface 595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f; detaching it from the instance and deleting it from the info cache [ 816.762256] env[63175]: DEBUG nova.network.neutron [req-b316d997-a552-4f9a-ab8e-5c454633100d req-1bcac634-aae8-49e1-a582-4ceea647ef58 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.992048] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24e558dc-d185-44df-89f9-4cf743b168dd tempest-InstanceActionsNegativeTestJSON-1114257627 tempest-InstanceActionsNegativeTestJSON-1114257627-project-member] Lock "39e55060-73a7-4dbd-96cf-bc48d8737c1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.597s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.189945] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248084, 'name': CreateVM_Task, 'duration_secs': 0.414847} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.189945] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.191485] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.191674] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.192022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.192303] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0354943b-b968-4947-9492-576920e56f23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.197335] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 817.197335] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ca5a5-8aad-c658-f5d1-f8ff4f07572b" [ 817.197335] env[63175]: _type = "Task" [ 817.197335] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.205920] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ca5a5-8aad-c658-f5d1-f8ff4f07572b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.232966] env[63175]: DEBUG nova.network.neutron [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.269223] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b196092-3ae3-496b-95bd-4f740ac93e04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.277093] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e833f0a-e711-43b2-90ea-84bbbe17f3c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.305844] env[63175]: DEBUG nova.compute.manager [req-b316d997-a552-4f9a-ab8e-5c454633100d req-1bcac634-aae8-49e1-a582-4ceea647ef58 service nova] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Detach interface failed, port_id=595cdfbb-7c3a-44ba-8479-bfb3a9ecbe8f, reason: Instance f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 817.489829] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Successfully updated port: 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.519598] env[63175]: DEBUG nova.network.neutron [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Updated VIF entry in instance network info cache for port 81942ffb-5329-4928-a222-00d5cb4e5d43. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.519776] env[63175]: DEBUG nova.network.neutron [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Updating instance_info_cache with network_info: [{"id": "81942ffb-5329-4928-a222-00d5cb4e5d43", "address": "fa:16:3e:79:33:fc", "network": {"id": "91cb4fda-fa7b-4ceb-abb1-553c86697f9a", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-50944530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "afe51efa241a40a5ae4c306b1688ccae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23f4655e-3495-421d-be4e-f6002a85a47a", "external-id": "nsx-vlan-transportzone-520", "segmentation_id": 520, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81942ffb-53", "ovs_interfaceid": "81942ffb-5329-4928-a222-00d5cb4e5d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.714503] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ca5a5-8aad-c658-f5d1-f8ff4f07572b, 'name': SearchDatastore_Task, 'duration_secs': 0.01085} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.717445] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.717689] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.717925] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.718089] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.718271] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.719023] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e417b6b7-b717-4e30-b43a-3f24c37a2aa7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.727684] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.727684] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.728816] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-587fa846-5205-4ada-b0f2-37ace23449c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.734764] env[63175]: INFO nova.compute.manager [-] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Took 1.35 seconds to deallocate network for instance. [ 817.737960] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 817.737960] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52522dc1-c809-b698-a9d0-34e43654d9d0" [ 817.737960] env[63175]: _type = "Task" [ 817.737960] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.750876] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52522dc1-c809-b698-a9d0-34e43654d9d0, 'name': SearchDatastore_Task, 'duration_secs': 0.009033} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.751829] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7570e541-5d1b-45a2-9f74-6ce0f9630ec2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.764381] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 817.764381] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5230bba8-4445-453c-26bc-2248b0786b00" [ 817.764381] env[63175]: _type = "Task" [ 817.764381] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.772237] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5230bba8-4445-453c-26bc-2248b0786b00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.774927] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6a872d-f03b-486a-8e0e-a477fdade1ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.783306] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce611c0-d7bf-4509-b90e-5985cea63aec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.815744] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bdeac6-2bf1-4a8f-ad55-2cf4d9dd8e22 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.823371] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb9b4d7-e80d-4994-826e-3234b54e8bcf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.838456] env[63175]: DEBUG nova.compute.provider_tree [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.995796] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.995979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.996412] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.023850] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ebb630e-d165-4093-94bd-e2566ffc770b req-0f412a5b-8a70-4bf1-86a2-fc02b96ef7f6 service nova] Releasing lock "refresh_cache-cf12aac3-edbb-48eb-b431-70187a4ecda3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.245772] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.275657] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5230bba8-4445-453c-26bc-2248b0786b00, 'name': SearchDatastore_Task, 'duration_secs': 0.009251} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.276098] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.276219] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] cf12aac3-edbb-48eb-b431-70187a4ecda3/cf12aac3-edbb-48eb-b431-70187a4ecda3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.276463] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82bd0550-58bb-476f-b394-c3871c6b2c4e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.283341] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 818.283341] env[63175]: value = "task-1248085" [ 818.283341] env[63175]: _type = "Task" [ 818.283341] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.291605] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.341866] env[63175]: DEBUG nova.scheduler.client.report [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 818.551233] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.753128] env[63175]: DEBUG nova.network.neutron [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.793315] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248085, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446108} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.794893] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] cf12aac3-edbb-48eb-b431-70187a4ecda3/cf12aac3-edbb-48eb-b431-70187a4ecda3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.794893] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.797032] env[63175]: DEBUG nova.compute.manager [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 818.797032] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.797032] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.797032] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.797032] env[63175]: DEBUG nova.compute.manager [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] No waiting events found dispatching network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 818.798258] env[63175]: WARNING nova.compute.manager [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received unexpected event network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb for instance with vm_state building and task_state spawning. [ 818.798258] env[63175]: DEBUG nova.compute.manager [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 818.798258] env[63175]: DEBUG nova.compute.manager [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing instance network info cache due to event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 818.798258] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.798258] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da170ff5-d814-4e6b-be09-fea0fb323fa6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.804277] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 818.804277] env[63175]: value = "task-1248086" [ 818.804277] env[63175]: _type = "Task" [ 818.804277] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.812429] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.854095] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.854623] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 818.857651] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.012s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.857867] env[63175]: DEBUG nova.objects.instance [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lazy-loading 'resources' on Instance uuid 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.244827] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.245103] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.257469] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.257469] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance network_info: |[{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 819.257852] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.257852] env[63175]: DEBUG nova.network.neutron [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.257852] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:c0:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a69f7df-439f-4044-8394-305f9f3a93bb', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.266221] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating folder: Project (5ea6da3858c347d887dcfba2556fabad). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.267711] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0beafdf-b92b-4b40-b1ef-314923471c11 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.279058] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Created folder: Project (5ea6da3858c347d887dcfba2556fabad) in parent group-v268956. [ 819.279264] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating folder: Instances. Parent ref: group-v269022. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.279605] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7891d255-cee6-4754-b868-8e15c94a18ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.289034] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Created folder: Instances in parent group-v269022. [ 819.289272] env[63175]: DEBUG oslo.service.loopingcall [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.289559] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.289788] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2cbab44-f50d-4353-b5f0-452a960dc284 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.316969] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065968} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.318576] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.318846] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.318846] env[63175]: value = "task-1248089" [ 819.318846] env[63175]: _type = "Task" [ 819.318846] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.319522] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec47708-9bbf-4142-b5ec-5a1b1e26b531 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.329381] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248089, 'name': CreateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.347069] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] cf12aac3-edbb-48eb-b431-70187a4ecda3/cf12aac3-edbb-48eb-b431-70187a4ecda3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.348315] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a18871fd-8664-436f-adc4-2161d080b7e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.365737] env[63175]: DEBUG nova.compute.utils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.370070] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 819.370422] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.379096] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 819.379096] env[63175]: value = "task-1248090" [ 819.379096] env[63175]: _type = "Task" [ 819.379096] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.387107] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248090, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.421879] env[63175]: DEBUG nova.policy [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '782ebbe4b1104fe68f7f5dcd35854a33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975415f67c2645678f0815424128063b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.664596] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea474218-5a89-4a82-9b28-6ed3f70b6690 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.672951] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75b4363-5f04-48cf-b2d0-a2b9a3b83684 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.702916] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af1213a-6ee4-44bb-9940-31d1d07b55c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.710843] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0acfea6b-e50b-4dc2-9c83-4136c43094c6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.718764] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully created port: 9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.728830] env[63175]: DEBUG nova.compute.provider_tree [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.747518] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 819.833407] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248089, 'name': CreateVM_Task, 'duration_secs': 0.469431} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.833881] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.834581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.834785] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.835087] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.835411] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad86715b-82f7-41d0-9760-57d95a7acb28 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.840546] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 819.840546] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52798a16-0b66-5c51-d5f9-27cec5b74c33" [ 819.840546] env[63175]: _type = "Task" [ 819.840546] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.848941] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52798a16-0b66-5c51-d5f9-27cec5b74c33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.870517] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 819.890720] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248090, 'name': ReconfigVM_Task, 'duration_secs': 0.474325} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.890952] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Reconfigured VM instance instance-0000003f to attach disk [datastore2] cf12aac3-edbb-48eb-b431-70187a4ecda3/cf12aac3-edbb-48eb-b431-70187a4ecda3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.891872] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11393964-cd6e-41c6-a287-169102cedbfc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.898361] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 819.898361] env[63175]: value = "task-1248091" [ 819.898361] env[63175]: _type = "Task" [ 819.898361] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.906415] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248091, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.021934] env[63175]: DEBUG nova.network.neutron [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updated VIF entry in instance network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.022382] env[63175]: DEBUG nova.network.neutron [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.121333] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully created port: d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.232275] env[63175]: DEBUG nova.scheduler.client.report [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 820.272986] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.351345] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52798a16-0b66-5c51-d5f9-27cec5b74c33, 'name': SearchDatastore_Task, 'duration_secs': 0.009645} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.351635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.351869] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.352109] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.352287] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.352474] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.352812] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9b3d5da-9b97-4570-a6bd-3e4ea6731885 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.364191] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.364409] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.365165] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-998380d8-f120-4182-b443-e6e20f71edf5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.371598] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 820.371598] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5242bdfb-6fca-4376-d52c-c6c9bf011a5e" [ 820.371598] env[63175]: _type = "Task" [ 820.371598] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.385054] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5242bdfb-6fca-4376-d52c-c6c9bf011a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.008007} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.385718] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-861578e0-3b5d-4104-a9f8-9a76a6120628 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.390734] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 820.390734] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fb9bf3-961a-87aa-e3c7-d9269bbe37ef" [ 820.390734] env[63175]: _type = "Task" [ 820.390734] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.394806] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully created port: 9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.401667] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fb9bf3-961a-87aa-e3c7-d9269bbe37ef, 'name': SearchDatastore_Task, 'duration_secs': 0.007569} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.404734] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.404990] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.405597] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b99e085-3ce1-475b-bcce-66b38d7c65a5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.413988] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248091, 'name': Rename_Task, 'duration_secs': 0.138222} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.415799] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.416126] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 820.416126] env[63175]: value = "task-1248092" [ 820.416126] env[63175]: _type = "Task" [ 820.416126] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.416316] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d1040a9-de9a-440d-b099-864232299fad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.426500] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248092, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.428097] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 820.428097] env[63175]: value = "task-1248093" [ 820.428097] env[63175]: _type = "Task" [ 820.428097] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.436249] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.525606] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e550f84-cfc6-4a7f-aabd-7b0c0df08ef6 req-47611927-0d2f-4e2d-870a-791d1d927be4 service nova] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.738143] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.741212] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.234s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.744266] env[63175]: INFO nova.compute.claims [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.778403] env[63175]: INFO nova.scheduler.client.report [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Deleted allocations for instance 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4 [ 820.881396] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 820.915220] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.915540] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.915721] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.915980] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.916187] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.916363] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.916596] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.916764] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.916959] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.917164] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.917350] env[63175]: DEBUG nova.virt.hardware [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.918670] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb187f8-5865-4813-9a9b-b40d46981cb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.940661] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81aefe1-aecc-4279-bb11-6c95b8c68149 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.944995] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248092, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47603} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.945309] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.945485] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.946413] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e249b0c-e863-40c5-b165-85631821b088 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.960627] env[63175]: DEBUG oslo_vmware.api [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248093, 'name': PowerOnVM_Task, 'duration_secs': 0.509861} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.962446] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.962686] env[63175]: INFO nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Took 6.82 seconds to spawn the instance on the hypervisor. [ 820.962866] env[63175]: DEBUG nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 820.963636] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015a277d-a78a-499e-97d0-7f2d7d840407 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.976136] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 820.976136] env[63175]: value = "task-1248094" [ 820.976136] env[63175]: _type = "Task" [ 820.976136] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.984439] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.286889] env[63175]: DEBUG oslo_concurrency.lockutils [None req-576d758a-2106-485c-81b0-9680554cc7d3 tempest-ServersAaction247Test-1296579289 tempest-ServersAaction247Test-1296579289-project-member] Lock "35f775ef-8a9e-4c49-99b9-a90f8c8f39b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.764s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.501678] env[63175]: INFO nova.compute.manager [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Took 31.38 seconds to build instance. [ 821.503102] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123508} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.503357] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.504273] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12af1e9-1eb3-4fa2-a000-7fb234f1e2ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.529858] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.530761] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddaa9ce8-7344-48e8-8f46-a6d602e05272 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.553164] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 821.553164] env[63175]: value = "task-1248095" [ 821.553164] env[63175]: _type = "Task" [ 821.553164] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.560397] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248095, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.003865] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0c8f1553-d025-4b35-aac4-5662b80d8082 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.499s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.047312] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a713470-b3ba-413b-9465-c6fd6d1ed0d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.058345] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f4d93e-b76b-426d-8a17-902df9115f81 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.065054] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248095, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.103011] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75fe59a-f205-4eac-bbf7-e3f33e4e4110 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.112548] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c3fa16-ce94-4a99-8c69-18a869397fd6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.119406] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "780989ee-98a1-4e99-9014-45de6921d4bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.119650] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.120040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.120277] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.120455] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.123251] env[63175]: INFO nova.compute.manager [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Terminating instance [ 822.133721] env[63175]: DEBUG nova.compute.provider_tree [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.191594] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.192065] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.192065] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.192603] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.192603] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.194616] env[63175]: INFO nova.compute.manager [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Terminating instance [ 822.325160] env[63175]: DEBUG nova.compute.manager [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-plugged-9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 822.325588] env[63175]: DEBUG oslo_concurrency.lockutils [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.325932] env[63175]: DEBUG oslo_concurrency.lockutils [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.326152] env[63175]: DEBUG oslo_concurrency.lockutils [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.326345] env[63175]: DEBUG nova.compute.manager [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] No waiting events found dispatching network-vif-plugged-9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 822.326519] env[63175]: WARNING nova.compute.manager [req-535488f1-968d-451b-b365-ae24220c8106 req-1481ff87-8912-4900-9e91-b0c94acda02e service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received unexpected event network-vif-plugged-9a593628-832c-4d9f-98f5-b300891488d8 for instance with vm_state building and task_state spawning. [ 822.426483] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully updated port: 9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.566024] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248095, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.639621] env[63175]: DEBUG nova.scheduler.client.report [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 822.645626] env[63175]: DEBUG nova.compute.manager [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 822.645626] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.647180] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1591ede1-5d29-44d5-a53f-27ce996610bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.659801] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.662430] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1196bf9-a25a-4315-8335-5a33a63ff9dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.669326] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 822.669326] env[63175]: value = "task-1248096" [ 822.669326] env[63175]: _type = "Task" [ 822.669326] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.678886] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.701863] env[63175]: DEBUG nova.compute.manager [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 822.702132] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.703458] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51cb8da-68fb-4d25-b84a-0cd850a25245 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.714479] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.714479] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd2360a2-61b6-41f3-bab1-a239d443464c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.718479] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 822.718479] env[63175]: value = "task-1248097" [ 822.718479] env[63175]: _type = "Task" [ 822.718479] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.727021] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248097, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.063673] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248095, 'name': ReconfigVM_Task, 'duration_secs': 1.292761} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.063962] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.064612] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd7115b3-ff6e-475f-972a-caefa283cde4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.070842] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 823.070842] env[63175]: value = "task-1248098" [ 823.070842] env[63175]: _type = "Task" [ 823.070842] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.078799] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248098, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.145970] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.146661] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 823.149679] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.930s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.152091] env[63175]: INFO nova.compute.claims [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.184362] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248096, 'name': PowerOffVM_Task, 'duration_secs': 0.217195} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.185311] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 823.185311] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.185311] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b4eca49-d7bc-4edd-a286-0403be0138bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.231100] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248097, 'name': PowerOffVM_Task, 'duration_secs': 0.174424} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.232824] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 823.232824] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.232824] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-198ab2d1-46ba-4dbb-9d6e-2b47b7512988 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.258369] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.258659] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.258884] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleting the datastore file [datastore2] 780989ee-98a1-4e99-9014-45de6921d4bc {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.260284] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-322dc73c-8d64-49aa-a487-25839457b1b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.266602] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 823.266602] env[63175]: value = "task-1248101" [ 823.266602] env[63175]: _type = "Task" [ 823.266602] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.280420] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.313528] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.313799] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.313991] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleting the datastore file [datastore2] cf12aac3-edbb-48eb-b431-70187a4ecda3 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.314287] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-853c35f1-a663-4add-901f-973e8b7dcc98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.322639] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for the task: (returnval){ [ 823.322639] env[63175]: value = "task-1248102" [ 823.322639] env[63175]: _type = "Task" [ 823.322639] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.330757] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248102, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.581551] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248098, 'name': Rename_Task, 'duration_secs': 0.192189} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.581856] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.582120] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b4af1af-63db-4665-af0b-37c94c93292d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.588494] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 823.588494] env[63175]: value = "task-1248103" [ 823.588494] env[63175]: _type = "Task" [ 823.588494] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.598633] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.659569] env[63175]: DEBUG nova.compute.utils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.660601] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 823.660755] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.707622] env[63175]: DEBUG nova.policy [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1502badb089644d4a06e6bec34e45a8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822ae9f380c84b3c914770f1232975b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.776105] env[63175]: DEBUG oslo_vmware.api [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154615} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.776916] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.776916] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.776916] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.776916] env[63175]: INFO nova.compute.manager [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Took 1.13 seconds to destroy the instance on the hypervisor. [ 823.777380] env[63175]: DEBUG oslo.service.loopingcall [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.777380] env[63175]: DEBUG nova.compute.manager [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 823.777445] env[63175]: DEBUG nova.network.neutron [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.834021] env[63175]: DEBUG oslo_vmware.api [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Task: {'id': task-1248102, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154078} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.834309] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.834618] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.834839] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.835059] env[63175]: INFO nova.compute.manager [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 823.835317] env[63175]: DEBUG oslo.service.loopingcall [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.835597] env[63175]: DEBUG nova.compute.manager [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 823.835703] env[63175]: DEBUG nova.network.neutron [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.031848] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Successfully created port: 25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.103218] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248103, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.164225] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 824.362184] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-changed-9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.362513] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing instance network info cache due to event network-changed-9a593628-832c-4d9f-98f5-b300891488d8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 824.362693] env[63175]: DEBUG oslo_concurrency.lockutils [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] Acquiring lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.362853] env[63175]: DEBUG oslo_concurrency.lockutils [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] Acquired lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.363062] env[63175]: DEBUG nova.network.neutron [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing network info cache for port 9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.394017] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.394244] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.447361] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c668a146-10cf-4713-b360-3fd7b07b17f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.456159] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b984b07-c4ec-453f-9786-d24139cf87f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.488749] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8768ec-3614-4703-b380-824c89166cd1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.496476] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06407442-19b9-4389-a341-dad1de9802b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.508932] env[63175]: DEBUG nova.compute.provider_tree [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.598394] env[63175]: DEBUG oslo_vmware.api [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248103, 'name': PowerOnVM_Task, 'duration_secs': 0.597347} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.598673] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.598880] env[63175]: INFO nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Took 8.00 seconds to spawn the instance on the hypervisor. [ 824.599122] env[63175]: DEBUG nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 824.599932] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a00313-f852-4aac-a774-c21ccf7dd872 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.659549] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully updated port: d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.717441] env[63175]: DEBUG nova.network.neutron [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.722030] env[63175]: DEBUG nova.network.neutron [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.896752] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 824.942556] env[63175]: DEBUG nova.network.neutron [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.012202] env[63175]: DEBUG nova.scheduler.client.report [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 825.030708] env[63175]: DEBUG nova.network.neutron [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.117118] env[63175]: INFO nova.compute.manager [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Took 32.63 seconds to build instance. [ 825.179195] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 825.209825] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.209825] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.209825] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.210068] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.210068] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.210068] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.210068] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.210068] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.210277] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.210314] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.210503] env[63175]: DEBUG nova.virt.hardware [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.211309] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd21fa32-5a37-4c4e-9b51-d600943e5353 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.219632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3ccf21-d072-4454-8ab4-60ddd4a7e99e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.224308] env[63175]: INFO nova.compute.manager [-] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Took 1.45 seconds to deallocate network for instance. [ 825.224558] env[63175]: INFO nova.compute.manager [-] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Took 1.39 seconds to deallocate network for instance. [ 825.361754] env[63175]: DEBUG nova.compute.manager [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 825.361953] env[63175]: DEBUG nova.compute.manager [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing instance network info cache due to event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 825.362190] env[63175]: DEBUG oslo_concurrency.lockutils [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.362432] env[63175]: DEBUG oslo_concurrency.lockutils [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.362525] env[63175]: DEBUG nova.network.neutron [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.425527] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.518348] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.518348] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 825.521335] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.640s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.521581] env[63175]: DEBUG nova.objects.instance [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lazy-loading 'resources' on Instance uuid e137904f-dc43-4ebb-90ab-e10ea5487fe5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.532872] env[63175]: DEBUG oslo_concurrency.lockutils [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] Releasing lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.533122] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Received event network-vif-deleted-1a07e685-ea5f-4104-8423-b0a5b0aa6a82 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 825.533297] env[63175]: INFO nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Neutron deleted interface 1a07e685-ea5f-4104-8423-b0a5b0aa6a82; detaching it from the instance and deleting it from the info cache [ 825.533485] env[63175]: DEBUG nova.network.neutron [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.619453] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cfdf7bd9-f4e6-465e-8850-9181eafa2063 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.747s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.673705] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Successfully updated port: 25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.735882] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.746887] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.026149] env[63175]: DEBUG nova.compute.utils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.030961] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 826.030961] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.036743] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c2d6811-1412-4418-a79c-4ac7e1d0e39c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.050079] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38a2edd-60e5-4a7f-95cd-653aa4dab9bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.087943] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Detach interface failed, port_id=1a07e685-ea5f-4104-8423-b0a5b0aa6a82, reason: Instance 780989ee-98a1-4e99-9014-45de6921d4bc could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 826.088188] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Received event network-vif-deleted-81942ffb-5329-4928-a222-00d5cb4e5d43 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 826.088367] env[63175]: INFO nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Neutron deleted interface 81942ffb-5329-4928-a222-00d5cb4e5d43; detaching it from the instance and deleting it from the info cache [ 826.088571] env[63175]: DEBUG nova.network.neutron [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.093870] env[63175]: DEBUG nova.policy [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1502badb089644d4a06e6bec34e45a8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822ae9f380c84b3c914770f1232975b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.179759] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.179946] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.180559] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.357065] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2c898f-05d2-432f-a027-2d8c2dd46f71 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.369301] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33606b1d-f6da-4104-b971-410a290a95fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.400243] env[63175]: DEBUG nova.network.neutron [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updated VIF entry in instance network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.400623] env[63175]: DEBUG nova.network.neutron [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.402468] env[63175]: DEBUG nova.compute.manager [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-plugged-d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 826.402680] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.402880] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.403058] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.403223] env[63175]: DEBUG nova.compute.manager [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] No waiting events found dispatching network-vif-plugged-d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 826.403387] env[63175]: WARNING nova.compute.manager [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received unexpected event network-vif-plugged-d266b161-204b-4b1c-b255-0263c31dfa40 for instance with vm_state building and task_state spawning. [ 826.403547] env[63175]: DEBUG nova.compute.manager [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-changed-d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 826.403697] env[63175]: DEBUG nova.compute.manager [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing instance network info cache due to event network-changed-d266b161-204b-4b1c-b255-0263c31dfa40. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 826.403873] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Acquiring lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.404016] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Acquired lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.404173] env[63175]: DEBUG nova.network.neutron [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing network info cache for port d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.405504] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc278c2d-0d24-4774-8f34-307a33a2b6b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.414789] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b85545-2f3a-47e7-a35e-7c1475ad3e1e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.421387] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Successfully created port: 8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.434642] env[63175]: DEBUG nova.compute.provider_tree [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.531239] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 826.591253] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c74f5d25-cc56-4de4-9740-19143f8c07c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.600474] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4f01f9-f95e-4846-887b-ee0aa82e4ba5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.628583] env[63175]: DEBUG nova.compute.manager [req-5786d20f-aca5-438f-8133-069222a12dfc req-4a816f4e-3023-4a40-9864-3091db5970aa service nova] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Detach interface failed, port_id=81942ffb-5329-4928-a222-00d5cb4e5d43, reason: Instance cf12aac3-edbb-48eb-b431-70187a4ecda3 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 826.730963] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.905513] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Successfully updated port: 9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.909493] env[63175]: DEBUG oslo_concurrency.lockutils [req-abf30848-763c-4295-a5a7-aa43ef3a0d05 req-533dcb9a-0713-4919-8a0c-d043bacce6fe service nova] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.940018] env[63175]: DEBUG nova.scheduler.client.report [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 826.948236] env[63175]: DEBUG nova.network.neutron [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.952893] env[63175]: DEBUG nova.network.neutron [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.047927] env[63175]: DEBUG nova.network.neutron [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.383565] env[63175]: DEBUG nova.compute.manager [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Received event network-vif-plugged-25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 827.383804] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Acquiring lock "a460926e-9637-40aa-bb30-e3890a441e03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.383983] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Lock "a460926e-9637-40aa-bb30-e3890a441e03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.384146] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Lock "a460926e-9637-40aa-bb30-e3890a441e03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.384326] env[63175]: DEBUG nova.compute.manager [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] No waiting events found dispatching network-vif-plugged-25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 827.384492] env[63175]: WARNING nova.compute.manager [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Received unexpected event network-vif-plugged-25225abe-2dcf-4815-85da-5fbf5ea4f2b5 for instance with vm_state building and task_state spawning. [ 827.384671] env[63175]: DEBUG nova.compute.manager [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Received event network-changed-25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 827.384789] env[63175]: DEBUG nova.compute.manager [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Refreshing instance network info cache due to event network-changed-25225abe-2dcf-4815-85da-5fbf5ea4f2b5. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 827.384954] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Acquiring lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.406835] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.448256] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.927s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.451069] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.618s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.451069] env[63175]: DEBUG nova.objects.instance [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lazy-loading 'resources' on Instance uuid a2e70590-17ed-4804-b232-57526e87d22b {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.455846] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.456140] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Instance network_info: |[{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 827.456429] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Acquired lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.456580] env[63175]: DEBUG nova.network.neutron [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Refreshing network info cache for port 25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.457695] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:57:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0467beaa-08c6-44d6-b8a2-e9c609c21ff4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25225abe-2dcf-4815-85da-5fbf5ea4f2b5', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.465827] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating folder: Project (822ae9f380c84b3c914770f1232975b7). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.468994] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e662a271-24cd-40c8-838f-a2360cc66bbe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.471600] env[63175]: INFO nova.scheduler.client.report [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Deleted allocations for instance e137904f-dc43-4ebb-90ab-e10ea5487fe5 [ 827.484840] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created folder: Project (822ae9f380c84b3c914770f1232975b7) in parent group-v268956. [ 827.484840] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating folder: Instances. Parent ref: group-v269025. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.484840] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a18112a-e916-4682-a66b-1a0ac74c4cb2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.493296] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created folder: Instances in parent group-v269025. [ 827.493521] env[63175]: DEBUG oslo.service.loopingcall [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.493702] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.493897] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a070de6-f771-45fa-abb3-894d31a65eb6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.514700] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.514700] env[63175]: value = "task-1248106" [ 827.514700] env[63175]: _type = "Task" [ 827.514700] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.523104] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248106, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.539982] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 827.550302] env[63175]: DEBUG oslo_concurrency.lockutils [req-832e46a5-e297-452e-8815-b8e65e6b34dd req-aeb1955c-77c2-4c2d-9b6e-19271161010b service nova] Releasing lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.550437] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.550604] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.565123] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.565254] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.565300] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.565479] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.565626] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.565773] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.566479] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.566479] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.566479] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.566479] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.566663] env[63175]: DEBUG nova.virt.hardware [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.567770] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0733487-58b8-4f3c-ad2c-8bf0e1f47ff5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.578219] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c7ca9c-2bbb-4aba-815f-06f634ac49e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.738958] env[63175]: DEBUG nova.network.neutron [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updated VIF entry in instance network info cache for port 25225abe-2dcf-4815-85da-5fbf5ea4f2b5. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.739348] env[63175]: DEBUG nova.network.neutron [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.979237] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ce2cabf0-142c-4c9c-9e91-14463ae24280 tempest-ServersTestBootFromVolume-962300623 tempest-ServersTestBootFromVolume-962300623-project-member] Lock "e137904f-dc43-4ebb-90ab-e10ea5487fe5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.334s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.025529] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248106, 'name': CreateVM_Task, 'duration_secs': 0.333824} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.025529] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.025776] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.025937] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.029022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.029022] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfbfcd67-98f7-454a-9b93-f92d10a12283 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.031454] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 828.031454] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529b1bf5-5987-36b2-63fa-5d4357158829" [ 828.031454] env[63175]: _type = "Task" [ 828.031454] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.042546] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529b1bf5-5987-36b2-63fa-5d4357158829, 'name': SearchDatastore_Task, 'duration_secs': 0.009136} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.042836] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.043089] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.043326] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.043492] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.043669] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.044157] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a93ed940-b42c-4d8f-a250-5902c0d1f08f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.052027] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.052027] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.053761] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-494adaaa-0307-4835-ab16-418f0c8e24b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.059662] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 828.059662] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520a07c5-17d6-c302-945b-136a52ee7fc2" [ 828.059662] env[63175]: _type = "Task" [ 828.059662] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.074987] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520a07c5-17d6-c302-945b-136a52ee7fc2, 'name': SearchDatastore_Task, 'duration_secs': 0.008051} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.081018] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ad1d1f1-8628-41db-895d-8e34dc99a53b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.085650] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 828.085650] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f15f73-57ad-9d7e-1da0-99e126c9849e" [ 828.085650] env[63175]: _type = "Task" [ 828.085650] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.097803] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f15f73-57ad-9d7e-1da0-99e126c9849e, 'name': SearchDatastore_Task, 'duration_secs': 0.00783} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.098106] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.098414] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/a460926e-9637-40aa-bb30-e3890a441e03.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.098759] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93ddde6e-4bc0-48c7-932c-faccb38564fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.103660] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.110142] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 828.110142] env[63175]: value = "task-1248107" [ 828.110142] env[63175]: _type = "Task" [ 828.110142] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.120946] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248107, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.178842] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Successfully updated port: 8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.210380] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fdf08a-2171-4837-ab0a-a5f90360278e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.219118] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8c45da-b71f-4dbc-a7d9-b3996c07fed8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.251245] env[63175]: DEBUG oslo_concurrency.lockutils [req-62a5b53e-ca84-4ebd-bf65-decbe472e2a1 req-340eb3fc-9a3d-483b-9760-94b31385fd75 service nova] Releasing lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.254946] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1afe62-045a-4f7d-9e71-7c0f3e4a974c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.264247] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b375cf43-ab22-4b70-ad50-dbc6e15f1034 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.279545] env[63175]: DEBUG nova.compute.provider_tree [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.498950] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-plugged-9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.499897] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.500179] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.500355] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.500528] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] No waiting events found dispatching network-vif-plugged-9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 828.500771] env[63175]: WARNING nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received unexpected event network-vif-plugged-9a92e009-2ca8-4a49-b16c-5c86c8270a78 for instance with vm_state building and task_state spawning. [ 828.500947] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-changed-9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.501118] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing instance network info cache due to event network-changed-9a92e009-2ca8-4a49-b16c-5c86c8270a78. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 828.501291] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquiring lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.619958] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248107, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439947} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.620249] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/a460926e-9637-40aa-bb30-e3890a441e03.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.620466] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.622488] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b57f6fa-9abc-407b-a91a-cdeec46b8bf6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.629080] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 828.629080] env[63175]: value = "task-1248108" [ 828.629080] env[63175]: _type = "Task" [ 828.629080] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.637201] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248108, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.640360] env[63175]: DEBUG nova.network.neutron [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [{"id": "9a593628-832c-4d9f-98f5-b300891488d8", "address": "fa:16:3e:a3:54:10", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.97", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a593628-83", "ovs_interfaceid": "9a593628-832c-4d9f-98f5-b300891488d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d266b161-204b-4b1c-b255-0263c31dfa40", "address": "fa:16:3e:94:c5:87", "network": {"id": "de07eec4-03a9-450c-9075-40aac3bf8b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-50212911", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "750b5f9b-f78a-4650-9153-c5bb117e507c", "external-id": "nsx-vlan-transportzone-237", "segmentation_id": 237, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd266b161-20", "ovs_interfaceid": "d266b161-204b-4b1c-b255-0263c31dfa40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "address": "fa:16:3e:c6:97:7b", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a92e009-2c", "ovs_interfaceid": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.681754] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.681907] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.682346] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.783120] env[63175]: DEBUG nova.scheduler.client.report [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 829.138648] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248108, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059966} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.138933] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.139707] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb3ef9b-60f7-469a-8b9f-681d7033f535 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.142603] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.142985] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance network_info: |[{"id": "9a593628-832c-4d9f-98f5-b300891488d8", "address": "fa:16:3e:a3:54:10", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.97", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a593628-83", "ovs_interfaceid": "9a593628-832c-4d9f-98f5-b300891488d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d266b161-204b-4b1c-b255-0263c31dfa40", "address": "fa:16:3e:94:c5:87", "network": {"id": "de07eec4-03a9-450c-9075-40aac3bf8b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-50212911", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "750b5f9b-f78a-4650-9153-c5bb117e507c", "external-id": "nsx-vlan-transportzone-237", "segmentation_id": 237, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd266b161-20", "ovs_interfaceid": "d266b161-204b-4b1c-b255-0263c31dfa40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "address": "fa:16:3e:c6:97:7b", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a92e009-2c", "ovs_interfaceid": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 829.143269] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquired lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.143487] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Refreshing network info cache for port 9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.145689] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:54:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a593628-832c-4d9f-98f5-b300891488d8', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:c5:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '750b5f9b-f78a-4650-9153-c5bb117e507c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd266b161-204b-4b1c-b255-0263c31dfa40', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:97:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a92e009-2ca8-4a49-b16c-5c86c8270a78', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.159296] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Creating folder: Project (975415f67c2645678f0815424128063b). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.171845] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df111bed-298c-4a68-af25-dbad34e7c144 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.182061] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/a460926e-9637-40aa-bb30-e3890a441e03.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.182660] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c4f18a3-3fe0-43d0-a0fc-b38fa5cccdca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.204273] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 829.204273] env[63175]: value = "task-1248110" [ 829.204273] env[63175]: _type = "Task" [ 829.204273] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.208268] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Created folder: Project (975415f67c2645678f0815424128063b) in parent group-v268956. [ 829.208462] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Creating folder: Instances. Parent ref: group-v269028. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.209016] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4e5225c-f72b-4c22-94f0-26be6c2e68db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.215042] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.218250] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Created folder: Instances in parent group-v269028. [ 829.218421] env[63175]: DEBUG oslo.service.loopingcall [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.218679] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.218797] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dac2f4a-3e60-4e39-ac31-3886b33c18fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.240230] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.243219] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.243219] env[63175]: value = "task-1248112" [ 829.243219] env[63175]: _type = "Task" [ 829.243219] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.251784] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248112, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.287924] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.292360] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.652s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.292677] env[63175]: INFO nova.compute.claims [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.332110] env[63175]: INFO nova.scheduler.client.report [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Deleted allocations for instance a2e70590-17ed-4804-b232-57526e87d22b [ 829.406936] env[63175]: DEBUG nova.network.neutron [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updating instance_info_cache with network_info: [{"id": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "address": "fa:16:3e:b0:44:25", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec148c2-7c", "ovs_interfaceid": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.715540] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.758012] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248112, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.840249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b3bb597d-e677-42e2-aa8b-086ad698ff15 tempest-AttachInterfacesV270Test-2026494755 tempest-AttachInterfacesV270Test-2026494755-project-member] Lock "a2e70590-17ed-4804-b232-57526e87d22b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.602s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.909466] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.909798] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Instance network_info: |[{"id": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "address": "fa:16:3e:b0:44:25", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec148c2-7c", "ovs_interfaceid": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 829.910266] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:44:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0467beaa-08c6-44d6-b8a2-e9c609c21ff4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ec148c2-7cf8-4923-9c17-a6f88820cf10', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.918638] env[63175]: DEBUG oslo.service.loopingcall [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.920580] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.920840] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23e63f4e-be37-4861-8b81-995dc4beff2f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.945221] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.945221] env[63175]: value = "task-1248113" [ 829.945221] env[63175]: _type = "Task" [ 829.945221] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.952923] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248113, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.996832] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updated VIF entry in instance network info cache for port 9a92e009-2ca8-4a49-b16c-5c86c8270a78. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.997328] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [{"id": "9a593628-832c-4d9f-98f5-b300891488d8", "address": "fa:16:3e:a3:54:10", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.97", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a593628-83", "ovs_interfaceid": "9a593628-832c-4d9f-98f5-b300891488d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d266b161-204b-4b1c-b255-0263c31dfa40", "address": "fa:16:3e:94:c5:87", "network": {"id": "de07eec4-03a9-450c-9075-40aac3bf8b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-50212911", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "750b5f9b-f78a-4650-9153-c5bb117e507c", "external-id": "nsx-vlan-transportzone-237", "segmentation_id": 237, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd266b161-20", "ovs_interfaceid": "d266b161-204b-4b1c-b255-0263c31dfa40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "address": "fa:16:3e:c6:97:7b", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a92e009-2c", "ovs_interfaceid": "9a92e009-2ca8-4a49-b16c-5c86c8270a78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.217735] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248110, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.254498] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248112, 'name': CreateVM_Task, 'duration_secs': 0.988391} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.254689] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.255579] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.255740] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.256104] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.256836] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b37dc2f1-2c78-4d8d-8934-faa8073ed055 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.261140] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 830.261140] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52992f76-dbbe-49bf-cf04-cb88dfea6b18" [ 830.261140] env[63175]: _type = "Task" [ 830.261140] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.270471] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52992f76-dbbe-49bf-cf04-cb88dfea6b18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.458684] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248113, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.502577] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Releasing lock "refresh_cache-46baf816-eff2-400c-b81c-0d5f3ce8d01e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.503065] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Received event network-vif-plugged-8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 830.503437] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquiring lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.503781] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.504131] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.504481] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] No waiting events found dispatching network-vif-plugged-8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 830.504840] env[63175]: WARNING nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Received unexpected event network-vif-plugged-8ec148c2-7cf8-4923-9c17-a6f88820cf10 for instance with vm_state building and task_state spawning. [ 830.506019] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Received event network-changed-8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 830.506019] env[63175]: DEBUG nova.compute.manager [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Refreshing instance network info cache due to event network-changed-8ec148c2-7cf8-4923-9c17-a6f88820cf10. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 830.506019] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquiring lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.506019] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Acquired lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.506019] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Refreshing network info cache for port 8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.572015] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b1d7d3-9ba5-4399-b6f2-73ae76b44f7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.577238] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5d7454-35e4-4b0c-8540-178fc237da3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.614394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f318a218-299a-4186-8b55-c34744508779 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.623920] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53ba8c4-ddc0-439e-b1e3-c9c2fbe743f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.638363] env[63175]: DEBUG nova.compute.provider_tree [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.718925] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248110, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.772478] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52992f76-dbbe-49bf-cf04-cb88dfea6b18, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.772793] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.773049] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.773296] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.773475] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.773656] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.774135] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7cc78d5-5422-4276-aa6f-332b3f0f7a79 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.783201] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.783458] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.784115] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84465d15-12ec-4aa8-9017-fd8978478988 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.789112] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 830.789112] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d51e7-8c7e-6f78-2064-e5541495ff57" [ 830.789112] env[63175]: _type = "Task" [ 830.789112] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.797074] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d51e7-8c7e-6f78-2064-e5541495ff57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.956127] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248113, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.141878] env[63175]: DEBUG nova.scheduler.client.report [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 831.219606] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248110, 'name': ReconfigVM_Task, 'duration_secs': 1.777569} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.219922] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfigured VM instance instance-00000042 to attach disk [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/a460926e-9637-40aa-bb30-e3890a441e03.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.220581] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64edac52-dc95-4ad3-a437-e7be7153dbf3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.227382] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 831.227382] env[63175]: value = "task-1248114" [ 831.227382] env[63175]: _type = "Task" [ 831.227382] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.235858] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248114, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.246402] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.246691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.247205] env[63175]: DEBUG nova.objects.instance [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid c445245b-b7d3-49c6-82c5-1e8188c89b68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.300090] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523d51e7-8c7e-6f78-2064-e5541495ff57, 'name': SearchDatastore_Task, 'duration_secs': 0.008016} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.300821] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0fd847-da74-4c77-a5d9-71ad6a345197 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.306576] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 831.306576] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528e03a1-656b-6296-bfd9-fccbfb0ec4f9" [ 831.306576] env[63175]: _type = "Task" [ 831.306576] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.314871] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528e03a1-656b-6296-bfd9-fccbfb0ec4f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.337165] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updated VIF entry in instance network info cache for port 8ec148c2-7cf8-4923-9c17-a6f88820cf10. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.337534] env[63175]: DEBUG nova.network.neutron [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updating instance_info_cache with network_info: [{"id": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "address": "fa:16:3e:b0:44:25", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec148c2-7c", "ovs_interfaceid": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.459828] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248113, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.648153] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.648684] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 831.652845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.883s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.653096] env[63175]: DEBUG nova.objects.instance [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lazy-loading 'resources' on Instance uuid 72d131a0-2617-49a1-8aff-897908929bb0 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.741595] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248114, 'name': Rename_Task, 'duration_secs': 0.166603} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.741909] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.742128] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-162e0a27-d8fd-4e64-a72f-1a623789458c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.748033] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 831.748033] env[63175]: value = "task-1248115" [ 831.748033] env[63175]: _type = "Task" [ 831.748033] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.758191] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.821574] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528e03a1-656b-6296-bfd9-fccbfb0ec4f9, 'name': SearchDatastore_Task, 'duration_secs': 0.009686} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.821574] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.822287] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 46baf816-eff2-400c-b81c-0d5f3ce8d01e/46baf816-eff2-400c-b81c-0d5f3ce8d01e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.822573] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2581ef5-1000-4b37-9b15-3b74654d030c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.830537] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 831.830537] env[63175]: value = "task-1248116" [ 831.830537] env[63175]: _type = "Task" [ 831.830537] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.839602] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.840424] env[63175]: DEBUG nova.objects.instance [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid c445245b-b7d3-49c6-82c5-1e8188c89b68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.842013] env[63175]: DEBUG oslo_concurrency.lockutils [req-f5a7c75f-b0a4-42ae-81ab-b51af52e566c req-8bf23f3c-824d-4338-b3be-3cf309adfd22 service nova] Releasing lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.960037] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248113, 'name': CreateVM_Task, 'duration_secs': 1.646333} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.960270] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.961250] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.961413] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.961738] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.961997] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3763224e-5833-4731-a923-7253cc9ee176 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.967916] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 831.967916] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529930b8-ec3f-f9c6-70c5-9505ec938109" [ 831.967916] env[63175]: _type = "Task" [ 831.967916] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.976271] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529930b8-ec3f-f9c6-70c5-9505ec938109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.158963] env[63175]: DEBUG nova.compute.utils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.167249] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 832.167449] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.264126] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248115, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.318110] env[63175]: DEBUG nova.policy [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.344137] env[63175]: DEBUG nova.objects.base [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 832.344396] env[63175]: DEBUG nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.346127] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248116, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474442} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.346377] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 46baf816-eff2-400c-b81c-0d5f3ce8d01e/46baf816-eff2-400c-b81c-0d5f3ce8d01e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.346608] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.347105] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc02ea42-58bc-4727-9d29-3b022c7d0a43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.355211] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 832.355211] env[63175]: value = "task-1248117" [ 832.355211] env[63175]: _type = "Task" [ 832.355211] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.364668] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248117, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.407963] env[63175]: DEBUG nova.policy [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.482755] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529930b8-ec3f-f9c6-70c5-9505ec938109, 'name': SearchDatastore_Task, 'duration_secs': 0.009249} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.483747] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.483747] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.483979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.484136] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.484425] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.485954] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31c2ceb0-9134-462b-8cbf-657a3f7ec2da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.489499] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e74fff-bbb5-4d02-913e-5b391f5a2fb6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.497591] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85ac908-06b2-4579-a7a8-21d32865d143 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.502483] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.502483] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.503478] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aaffeed-effb-4fd0-ba9b-d126a822c799 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.533667] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1aa60f-75cf-4fe0-a5d4-6515722b7b6f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.537110] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 832.537110] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522fb141-856e-a2b0-fb34-dc57a34071cf" [ 832.537110] env[63175]: _type = "Task" [ 832.537110] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.542801] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82e7dd6-f77b-4045-a271-74d0386de20c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.549391] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522fb141-856e-a2b0-fb34-dc57a34071cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.559088] env[63175]: DEBUG nova.compute.provider_tree [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.668301] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 832.756960] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Successfully created port: 6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.766700] env[63175]: DEBUG oslo_vmware.api [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248115, 'name': PowerOnVM_Task, 'duration_secs': 0.735753} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.767140] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.767453] env[63175]: INFO nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Took 7.59 seconds to spawn the instance on the hypervisor. [ 832.767743] env[63175]: DEBUG nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.768783] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a9fbb1-b61a-40b9-8b81-843a4076e37f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.868447] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248117, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067057} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.873243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.873243] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96be31ab-d7f0-4ad7-abcc-5a10b39642a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.905969] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 46baf816-eff2-400c-b81c-0d5f3ce8d01e/46baf816-eff2-400c-b81c-0d5f3ce8d01e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.906463] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e70e8cf-ec36-42f2-a2c1-93df47697b43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.921517] env[63175]: DEBUG nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Successfully created port: 0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.929463] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 832.929463] env[63175]: value = "task-1248118" [ 832.929463] env[63175]: _type = "Task" [ 832.929463] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.939754] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248118, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.048819] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522fb141-856e-a2b0-fb34-dc57a34071cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01845} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.049765] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d61d2423-20f8-4169-ae70-9c0069db593a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.057553] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 833.057553] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f5f0cb-4111-bb48-8b8b-3a3208a3ca6f" [ 833.057553] env[63175]: _type = "Task" [ 833.057553] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.062211] env[63175]: DEBUG nova.scheduler.client.report [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 833.071789] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f5f0cb-4111-bb48-8b8b-3a3208a3ca6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.288517] env[63175]: INFO nova.compute.manager [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Took 33.80 seconds to build instance. [ 833.439887] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.569967] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.919s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.571968] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f5f0cb-4111-bb48-8b8b-3a3208a3ca6f, 'name': SearchDatastore_Task, 'duration_secs': 0.039489} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.572547] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.530s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.572651] env[63175]: DEBUG nova.objects.instance [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 833.575249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.575503] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/54440032-4d9b-41d4-9ef2-5a79a4224fa6.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.576302] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1df25e5-0a85-45cd-8306-43f240306c70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.582982] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 833.582982] env[63175]: value = "task-1248119" [ 833.582982] env[63175]: _type = "Task" [ 833.582982] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.591138] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248119, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.592199] env[63175]: INFO nova.scheduler.client.report [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Deleted allocations for instance 72d131a0-2617-49a1-8aff-897908929bb0 [ 833.683485] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 833.704704] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.704998] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.705193] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.705411] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.705567] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.705727] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.705971] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.706190] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.706410] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.706608] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.706800] env[63175]: DEBUG nova.virt.hardware [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.708027] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee95ec36-e307-468b-947f-4d1f3e95c05c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.716614] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9f506c-1bf8-4605-8b64-c941c7f3aeb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.791017] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e5c8e862-f13d-4017-bde9-4e0f00ea52f4 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.029s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.941595] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248118, 'name': ReconfigVM_Task, 'duration_secs': 0.870325} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.943160] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 46baf816-eff2-400c-b81c-0d5f3ce8d01e/46baf816-eff2-400c-b81c-0d5f3ce8d01e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.943160] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1725768a-7a74-453d-b264-daefb172cee5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.950611] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 833.950611] env[63175]: value = "task-1248120" [ 833.950611] env[63175]: _type = "Task" [ 833.950611] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.960411] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248120, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.000497] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.000798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.094732] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248119, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.103614] env[63175]: DEBUG oslo_concurrency.lockutils [None req-95172881-c291-45ed-ac59-e8930b3e1776 tempest-ServerAddressesNegativeTestJSON-1148597488 tempest-ServerAddressesNegativeTestJSON-1148597488-project-member] Lock "72d131a0-2617-49a1-8aff-897908929bb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.989s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.460549] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248120, 'name': Rename_Task, 'duration_secs': 0.302445} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.460823] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 834.461125] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f574151d-1501-4f00-a7a8-69f41c66c094 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.466975] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 834.466975] env[63175]: value = "task-1248121" [ 834.466975] env[63175]: _type = "Task" [ 834.466975] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.474393] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248121, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.504653] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 834.574820] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.575218] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.583354] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ff3ef0a3-1bb5-40f4-af34-4dbb1d2c7c37 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.584463] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.472s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.584597] env[63175]: DEBUG nova.objects.instance [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lazy-loading 'resources' on Instance uuid 4d07a537-7df6-4659-8760-bf7e7925da25 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.600806] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248119, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621379} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.602153] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/54440032-4d9b-41d4-9ef2-5a79a4224fa6.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.602311] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.602579] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-356ecf67-7667-4fa6-9e07-331987d5ab91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.609699] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 834.609699] env[63175]: value = "task-1248122" [ 834.609699] env[63175]: _type = "Task" [ 834.609699] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.619814] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.710652] env[63175]: DEBUG nova.compute.manager [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received event network-vif-plugged-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 834.710896] env[63175]: DEBUG oslo_concurrency.lockutils [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.711378] env[63175]: DEBUG oslo_concurrency.lockutils [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.711600] env[63175]: DEBUG oslo_concurrency.lockutils [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.711886] env[63175]: DEBUG nova.compute.manager [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] No waiting events found dispatching network-vif-plugged-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 834.712165] env[63175]: WARNING nova.compute.manager [req-27917a3e-2a60-4a99-bac9-9d207ece0ddf req-5a3f4560-82be-4634-89af-f1df467d8967 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received unexpected event network-vif-plugged-6f9757fd-9d26-48d8-89e3-3411a8770034 for instance with vm_state building and task_state spawning. [ 834.841188] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Successfully updated port: 6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.931196] env[63175]: DEBUG nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Successfully updated port: 0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.978704] env[63175]: DEBUG oslo_vmware.api [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248121, 'name': PowerOnVM_Task, 'duration_secs': 0.490372} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.978704] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.978978] env[63175]: INFO nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Took 14.10 seconds to spawn the instance on the hypervisor. [ 834.979144] env[63175]: DEBUG nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 834.980067] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425124af-8f6b-468e-a4cf-090d58d8d740 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.034827] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.077180] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 835.123706] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077518} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.124368] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.124840] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b19afc5-cbeb-46d9-8184-ab7aee5b8429 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.150348] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/54440032-4d9b-41d4-9ef2-5a79a4224fa6.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.153360] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5b17e23-8b56-4ba8-8ade-2cfec6025196 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.175389] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 835.175389] env[63175]: value = "task-1248123" [ 835.175389] env[63175]: _type = "Task" [ 835.175389] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.187080] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.343585] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.343979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.343979] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.354179] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d477e8-6aec-4a70-9d96-a3529fe15b98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.362295] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5420918-be04-4735-b68e-4b3ebd3381fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.396792] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8268860d-4217-4fb6-886b-c12f86f9fbca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.406392] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d04fa3-c829-4fb5-96d9-2b2a5b529893 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.420926] env[63175]: DEBUG nova.compute.provider_tree [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.434138] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.434331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.434511] env[63175]: DEBUG nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.508526] env[63175]: INFO nova.compute.manager [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Took 40.83 seconds to build instance. [ 835.606480] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.688090] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.892039] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.924248] env[63175]: DEBUG nova.scheduler.client.report [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 835.990187] env[63175]: WARNING nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 836.004525] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.009715] env[63175]: DEBUG oslo_concurrency.lockutils [None req-987c0cca-8ddb-43ee-abd9-f5ba6b4ef376 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.594s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.009977] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.006s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.010208] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.010508] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.010698] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.012976] env[63175]: INFO nova.compute.manager [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Terminating instance [ 836.130352] env[63175]: DEBUG nova.network.neutron [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updating instance_info_cache with network_info: [{"id": "6f9757fd-9d26-48d8-89e3-3411a8770034", "address": "fa:16:3e:ec:8a:6d", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f9757fd-9d", "ovs_interfaceid": "6f9757fd-9d26-48d8-89e3-3411a8770034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.185502] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248123, 'name': ReconfigVM_Task, 'duration_secs': 0.750154} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.185773] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/54440032-4d9b-41d4-9ef2-5a79a4224fa6.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.186412] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cfd62477-e46d-471e-8fb1-6d0bbd1c75d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.192358] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 836.192358] env[63175]: value = "task-1248124" [ 836.192358] env[63175]: _type = "Task" [ 836.192358] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.200083] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248124, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.341669] env[63175]: DEBUG nova.network.neutron [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "address": "fa:16:3e:21:6a:e6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e60107b-fb", "ovs_interfaceid": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.430973] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.433662] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.152s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.435146] env[63175]: INFO nova.compute.claims [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.454344] env[63175]: INFO nova.scheduler.client.report [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Deleted allocations for instance 4d07a537-7df6-4659-8760-bf7e7925da25 [ 836.516663] env[63175]: DEBUG nova.compute.manager [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 836.516945] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 836.517881] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db90f7c-5c5f-4da6-a0c4-e8bfd6c935a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.525600] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.525832] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc480090-9bcd-418e-a0ba-06f0c4450419 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.531457] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 836.531457] env[63175]: value = "task-1248125" [ 836.531457] env[63175]: _type = "Task" [ 836.531457] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.539319] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.633249] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.633578] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance network_info: |[{"id": "6f9757fd-9d26-48d8-89e3-3411a8770034", "address": "fa:16:3e:ec:8a:6d", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f9757fd-9d", "ovs_interfaceid": "6f9757fd-9d26-48d8-89e3-3411a8770034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 836.634047] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:8a:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f9757fd-9d26-48d8-89e3-3411a8770034', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.641812] env[63175]: DEBUG oslo.service.loopingcall [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.642081] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.642329] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-355b784b-d5fe-460e-8b6f-9e25f4d664b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.664174] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.664174] env[63175]: value = "task-1248126" [ 836.664174] env[63175]: _type = "Task" [ 836.664174] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.672814] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248126, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.702454] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248124, 'name': Rename_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.788506] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-vif-plugged-0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 836.788506] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.788506] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.788506] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.788506] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] No waiting events found dispatching network-vif-plugged-0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 836.788506] env[63175]: WARNING nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received unexpected event network-vif-plugged-0e60107b-fb39-48fa-8a14-dee31bc1bb00 for instance with vm_state active and task_state None. [ 836.788808] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received event network-changed-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 836.788808] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Refreshing instance network info cache due to event network-changed-6f9757fd-9d26-48d8-89e3-3411a8770034. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 836.789019] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Acquiring lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.789192] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Acquired lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.789394] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Refreshing network info cache for port 6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.845189] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.845880] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.846065] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.846882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4fbc71-24a7-4633-852b-1526b892b1bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.863997] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.864252] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.864412] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.864597] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.864745] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.864888] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.865110] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.865275] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.865440] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.865603] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.865772] env[63175]: DEBUG nova.virt.hardware [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.872350] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfiguring VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 836.873031] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f35ee96-e992-4bae-9d46-17e8f84a0838 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.889914] env[63175]: DEBUG oslo_vmware.api [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 836.889914] env[63175]: value = "task-1248127" [ 836.889914] env[63175]: _type = "Task" [ 836.889914] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.897962] env[63175]: DEBUG oslo_vmware.api [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248127, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.961430] env[63175]: DEBUG oslo_concurrency.lockutils [None req-74e52584-add5-441b-9192-fe06c78fe41f tempest-VolumesAdminNegativeTest-1962887060 tempest-VolumesAdminNegativeTest-1962887060-project-member] Lock "4d07a537-7df6-4659-8760-bf7e7925da25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.212s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.040743] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248125, 'name': PowerOffVM_Task, 'duration_secs': 0.188845} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.041015] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 837.041699] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 837.042090] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86f60c03-9190-4ba6-9761-e895e439b39f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.174084] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248126, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.201577] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248124, 'name': Rename_Task, 'duration_secs': 0.844761} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.201845] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.202105] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd732cb9-402a-49b9-9a4e-3af63bdcc205 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.208438] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 837.208438] env[63175]: value = "task-1248129" [ 837.208438] env[63175]: _type = "Task" [ 837.208438] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.224551] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.245657] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 837.245864] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 837.246071] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleting the datastore file [datastore1] 46baf816-eff2-400c-b81c-0d5f3ce8d01e {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.246356] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-91f3b469-f9ea-480c-9abb-8f80d1bcec73 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.252464] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 837.252464] env[63175]: value = "task-1248130" [ 837.252464] env[63175]: _type = "Task" [ 837.252464] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.262958] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248130, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.399909] env[63175]: DEBUG oslo_vmware.api [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248127, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.541041] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updated VIF entry in instance network info cache for port 6f9757fd-9d26-48d8-89e3-3411a8770034. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.541507] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updating instance_info_cache with network_info: [{"id": "6f9757fd-9d26-48d8-89e3-3411a8770034", "address": "fa:16:3e:ec:8a:6d", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f9757fd-9d", "ovs_interfaceid": "6f9757fd-9d26-48d8-89e3-3411a8770034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.677045] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248126, 'name': CreateVM_Task, 'duration_secs': 0.57801} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.677217] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.677894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.678550] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.678550] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.678669] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d009dbbb-7dbd-4a6d-b174-90594604310f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.681852] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f856c8-acf4-4533-b185-8850cd0cc728 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.687940] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 837.687940] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5265b7df-9209-92d2-3420-5e6c37733db9" [ 837.687940] env[63175]: _type = "Task" [ 837.687940] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.693208] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cd200b-e803-41ba-b6bc-99f3414d15b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.701934] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5265b7df-9209-92d2-3420-5e6c37733db9, 'name': SearchDatastore_Task, 'duration_secs': 0.010768} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.725781] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.726037] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.726277] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.726421] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.726597] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.727087] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34e3bd8e-63cc-4582-bae8-1951cb96c217 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.732034] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c803d2-8a9d-43ac-81b6-44f04f32e220 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.741403] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248129, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.743251] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb567475-768b-4aa1-97d6-f78d68806066 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.746917] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.747124] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.747775] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2f9b411-74df-4f23-9b85-bdc7a1b6d085 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.753060] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 837.753060] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52380289-59c1-8af6-6012-0778c7598b2b" [ 837.753060] env[63175]: _type = "Task" [ 837.753060] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.765441] env[63175]: DEBUG nova.compute.provider_tree [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.774719] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52380289-59c1-8af6-6012-0778c7598b2b, 'name': SearchDatastore_Task, 'duration_secs': 0.011434} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.778569] env[63175]: DEBUG oslo_vmware.api [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248130, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321966} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.778768] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8aefff2-1b7d-4698-b60d-525361980137 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.783428] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.783428] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 837.783428] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.783428] env[63175]: INFO nova.compute.manager [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Took 1.26 seconds to destroy the instance on the hypervisor. [ 837.783428] env[63175]: DEBUG oslo.service.loopingcall [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.783428] env[63175]: DEBUG nova.compute.manager [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 837.783428] env[63175]: DEBUG nova.network.neutron [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 837.787681] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 837.787681] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520664f3-0110-67c1-95d2-afa3ca514632" [ 837.787681] env[63175]: _type = "Task" [ 837.787681] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.796144] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520664f3-0110-67c1-95d2-afa3ca514632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.902373] env[63175]: DEBUG oslo_vmware.api [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248127, 'name': ReconfigVM_Task, 'duration_secs': 0.621037} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.902879] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.903112] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfigured VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 838.045422] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Releasing lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.045655] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-changed-0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 838.045827] env[63175]: DEBUG nova.compute.manager [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing instance network info cache due to event network-changed-0e60107b-fb39-48fa-8a14-dee31bc1bb00. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 838.046056] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.046205] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.046367] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Refreshing network info cache for port 0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.240171] env[63175]: DEBUG oslo_vmware.api [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248129, 'name': PowerOnVM_Task, 'duration_secs': 0.76892} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.240491] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.240712] env[63175]: INFO nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Took 10.70 seconds to spawn the instance on the hypervisor. [ 838.240887] env[63175]: DEBUG nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 838.241677] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2d8ea9-0fe8-4565-b741-11f0c2b001c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.270088] env[63175]: DEBUG nova.scheduler.client.report [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 838.299247] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520664f3-0110-67c1-95d2-afa3ca514632, 'name': SearchDatastore_Task, 'duration_secs': 0.015897} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.299463] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.299713] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] bfb59d40-c7f2-4f90-9387-8b886d2dbdba/bfb59d40-c7f2-4f90-9387-8b886d2dbdba.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.300327] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e68cc4a6-d3b0-4722-8f5b-515e38d08316 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.307734] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 838.307734] env[63175]: value = "task-1248131" [ 838.307734] env[63175]: _type = "Task" [ 838.307734] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.316410] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248131, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.407435] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e99effc5-14f3-4be8-ac85-efe8cf316629 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.161s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.766479] env[63175]: INFO nova.compute.manager [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Took 37.57 seconds to build instance. [ 838.777290] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.777754] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 838.779875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.534s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.780151] env[63175]: DEBUG nova.objects.instance [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'resources' on Instance uuid f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.821778] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248131, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.827902] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-deleted-9a92e009-2ca8-4a49-b16c-5c86c8270a78 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 838.828223] env[63175]: INFO nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Neutron deleted interface 9a92e009-2ca8-4a49-b16c-5c86c8270a78; detaching it from the instance and deleting it from the info cache [ 838.828962] env[63175]: DEBUG nova.network.neutron [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [{"id": "9a593628-832c-4d9f-98f5-b300891488d8", "address": "fa:16:3e:a3:54:10", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.97", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a593628-83", "ovs_interfaceid": "9a593628-832c-4d9f-98f5-b300891488d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d266b161-204b-4b1c-b255-0263c31dfa40", "address": "fa:16:3e:94:c5:87", "network": {"id": "de07eec4-03a9-450c-9075-40aac3bf8b17", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-50212911", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.239", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "750b5f9b-f78a-4650-9153-c5bb117e507c", "external-id": "nsx-vlan-transportzone-237", "segmentation_id": 237, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd266b161-20", "ovs_interfaceid": "d266b161-204b-4b1c-b255-0263c31dfa40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.832042] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updated VIF entry in instance network info cache for port 0e60107b-fb39-48fa-8a14-dee31bc1bb00. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.832337] env[63175]: DEBUG nova.network.neutron [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "address": "fa:16:3e:21:6a:e6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e60107b-fb", "ovs_interfaceid": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.933470] env[63175]: DEBUG nova.network.neutron [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.269631] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3e461088-3258-49c4-ba59-a5dd3d314c80 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.529s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.283444] env[63175]: DEBUG nova.compute.utils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.287729] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 839.287912] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.322071] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248131, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.812554} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.322398] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] bfb59d40-c7f2-4f90-9387-8b886d2dbdba/bfb59d40-c7f2-4f90-9387-8b886d2dbdba.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.322647] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.322919] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0aa69c7-7bb6-4b3f-8637-1f36e6b795b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.330510] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 839.330510] env[63175]: value = "task-1248132" [ 839.330510] env[63175]: _type = "Task" [ 839.330510] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.335792] env[63175]: DEBUG oslo_concurrency.lockutils [req-27f5aa42-057e-4a4f-a7df-d54da9a4d013 req-d9e05cf3-e9fa-4609-a205-7232b517f8a4 service nova] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.339399] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ab2b4c8-85b5-4fdf-bbb0-a6d3352ac843 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.341318] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.350675] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1ea913-8cd9-4adb-afaa-d6ed2f9da420 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.362992] env[63175]: DEBUG nova.policy [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c654f31fd4bc49dc8e985bd75a811ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74dd39e20aad4ddb874282973cea02bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.385791] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Detach interface failed, port_id=9a92e009-2ca8-4a49-b16c-5c86c8270a78, reason: Instance 46baf816-eff2-400c-b81c-0d5f3ce8d01e could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 839.386092] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-deleted-d266b161-204b-4b1c-b255-0263c31dfa40 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 839.386407] env[63175]: INFO nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Neutron deleted interface d266b161-204b-4b1c-b255-0263c31dfa40; detaching it from the instance and deleting it from the info cache [ 839.386618] env[63175]: DEBUG nova.network.neutron [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [{"id": "9a593628-832c-4d9f-98f5-b300891488d8", "address": "fa:16:3e:a3:54:10", "network": {"id": "d47e57c5-53b0-41ce-bb29-7a4ddf79605a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-970242295", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.97", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a593628-83", "ovs_interfaceid": "9a593628-832c-4d9f-98f5-b300891488d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.393946] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 839.395653] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 839.436748] env[63175]: INFO nova.compute.manager [-] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Took 1.65 seconds to deallocate network for instance. [ 839.564073] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32760d17-b6f9-48e4-b717-a1754251ca48 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.572320] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5735e4-ef82-4056-8be8-6ac1419187f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.618590] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6bbceb-4950-467a-90a1-509838ab69f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.626576] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cd81ae-c00d-4b5d-8ecf-6d0261984227 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.640775] env[63175]: DEBUG nova.compute.provider_tree [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.788855] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 839.843597] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061905} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.847108] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.847108] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decbfd68-1c6f-401d-b269-32bcd6570ec0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.876026] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] bfb59d40-c7f2-4f90-9387-8b886d2dbdba/bfb59d40-c7f2-4f90-9387-8b886d2dbdba.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.876026] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0c60406-6ecc-4845-b0e2-b3380ac0a656 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.897229] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36305520-de8e-4ed5-9ac4-245471065759 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.908470] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 839.908470] env[63175]: value = "task-1248133" [ 839.908470] env[63175]: _type = "Task" [ 839.908470] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.911903] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 839.912076] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 839.912259] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 839.916876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca59382f-086e-4e3a-a2dc-17e522208a1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.930591] env[63175]: INFO nova.compute.manager [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Rescuing [ 839.930847] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.930996] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.931173] env[63175]: DEBUG nova.network.neutron [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.936528] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248133, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.945647] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.956057] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Detach interface failed, port_id=d266b161-204b-4b1c-b255-0263c31dfa40, reason: Instance 46baf816-eff2-400c-b81c-0d5f3ce8d01e could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 839.956285] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Received event network-vif-deleted-9a593628-832c-4d9f-98f5-b300891488d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 839.956459] env[63175]: INFO nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Neutron deleted interface 9a593628-832c-4d9f-98f5-b300891488d8; detaching it from the instance and deleting it from the info cache [ 839.956629] env[63175]: DEBUG nova.network.neutron [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.958314] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Successfully created port: a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.143797] env[63175]: DEBUG nova.scheduler.client.report [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 840.416564] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.439458] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 840.439633] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Skipping network cache update for instance because it is Building. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 840.461337] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da6b6942-f500-4149-8791-42586908177a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.470600] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1834d6-0f74-4f4a-8ad8-f565b1a2a5a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.491156] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.491301] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.491444] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Forcefully refreshing network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 840.491604] env[63175]: DEBUG nova.objects.instance [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lazy-loading 'info_cache' on Instance uuid c445245b-b7d3-49c6-82c5-1e8188c89b68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.501756] env[63175]: DEBUG nova.compute.manager [req-5966a2dc-cf57-4f0d-92f1-e06b5d1eebb4 req-79c8db33-6af5-445e-804d-17b31ec816ff service nova] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Detach interface failed, port_id=9a593628-832c-4d9f-98f5-b300891488d8, reason: Instance 46baf816-eff2-400c-b81c-0d5f3ce8d01e could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 840.648791] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.650904] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.378s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.652817] env[63175]: INFO nova.compute.claims [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.669512] env[63175]: INFO nova.scheduler.client.report [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocations for instance f6d2c297-1cfe-4cab-8854-eb5760e6cbc0 [ 840.679497] env[63175]: DEBUG nova.network.neutron [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updating instance_info_cache with network_info: [{"id": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "address": "fa:16:3e:b0:44:25", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec148c2-7c", "ovs_interfaceid": "8ec148c2-7cf8-4923-9c17-a6f88820cf10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.760362] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-0e60107b-fb39-48fa-8a14-dee31bc1bb00" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.760652] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-0e60107b-fb39-48fa-8a14-dee31bc1bb00" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.801586] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 840.841206] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.841479] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.841622] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.841823] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.841982] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.842146] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.842706] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.842706] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.842819] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.843108] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.843250] env[63175]: DEBUG nova.virt.hardware [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.844087] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958e0100-74a3-476b-941b-6509c29e9ddb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.855046] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a694b536-81d8-46e5-af52-615208e07570 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.917846] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248133, 'name': ReconfigVM_Task, 'duration_secs': 0.918367} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.918246] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Reconfigured VM instance instance-00000044 to attach disk [datastore1] bfb59d40-c7f2-4f90-9387-8b886d2dbdba/bfb59d40-c7f2-4f90-9387-8b886d2dbdba.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.918899] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7e29f54-1765-49d5-86aa-dff09595a92c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.925959] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 840.925959] env[63175]: value = "task-1248134" [ 840.925959] env[63175]: _type = "Task" [ 840.925959] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.934461] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248134, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.176886] env[63175]: DEBUG oslo_concurrency.lockutils [None req-624bc354-0b9f-4c1d-b2d4-a163889eb8b4 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "f6d2c297-1cfe-4cab-8854-eb5760e6cbc0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.475s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.181501] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-54440032-4d9b-41d4-9ef2-5a79a4224fa6" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.263770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.263975] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.264892] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1f3765-ec01-4130-84ca-68ee3ff1280c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.282462] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4009dcb-1031-4db3-9249-bf4021496bfa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.308911] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfiguring VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 841.309236] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ac8ff5e-8e30-4f68-8779-b619de52bcc1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.327309] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 841.327309] env[63175]: value = "task-1248135" [ 841.327309] env[63175]: _type = "Task" [ 841.327309] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.335223] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.382889] env[63175]: DEBUG nova.compute.manager [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Received event network-vif-plugged-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 841.383121] env[63175]: DEBUG oslo_concurrency.lockutils [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.383421] env[63175]: DEBUG oslo_concurrency.lockutils [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.383618] env[63175]: DEBUG oslo_concurrency.lockutils [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.383769] env[63175]: DEBUG nova.compute.manager [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] No waiting events found dispatching network-vif-plugged-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 841.383940] env[63175]: WARNING nova.compute.manager [req-a3816170-6946-4d41-8623-7369626c6b65 req-cc4b016a-ced9-4239-a128-74a0906f9dbf service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Received unexpected event network-vif-plugged-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf for instance with vm_state building and task_state spawning. [ 841.438043] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248134, 'name': Rename_Task, 'duration_secs': 0.162562} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.438332] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.438591] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd9b04cf-b762-42f8-b42e-fed6aec68889 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.444499] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 841.444499] env[63175]: value = "task-1248136" [ 841.444499] env[63175]: _type = "Task" [ 841.444499] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.452109] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.500049] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Successfully updated port: a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.681372] env[63175]: DEBUG nova.scheduler.client.report [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Refreshing inventories for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 841.697475] env[63175]: DEBUG nova.scheduler.client.report [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Updating ProviderTree inventory for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 841.697760] env[63175]: DEBUG nova.compute.provider_tree [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.713127] env[63175]: DEBUG nova.scheduler.client.report [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Refreshing aggregate associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, aggregates: None {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 841.735102] env[63175]: DEBUG nova.scheduler.client.report [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Refreshing trait associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 841.840101] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.943381] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5469e45c-c651-479b-898c-72c297b7ad0e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.956872] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13e0b72-2e54-47e6-b50b-802ef18cc10f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.959968] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248136, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.989418] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10680a66-4903-4d73-bfce-366b4ba4ef1e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.998279] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bd7759-f0a0-47c2-9ab2-3667f7d5e5dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.002811] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.002983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.003111] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.014442] env[63175]: DEBUG nova.compute.provider_tree [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.074756] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.338833] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.360920] env[63175]: DEBUG nova.network.neutron [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.433631] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "address": "fa:16:3e:21:6a:e6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e60107b-fb", "ovs_interfaceid": "0e60107b-fb39-48fa-8a14-dee31bc1bb00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.456597] env[63175]: DEBUG oslo_vmware.api [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248136, 'name': PowerOnVM_Task, 'duration_secs': 0.808563} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.456945] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.457193] env[63175]: INFO nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Took 8.77 seconds to spawn the instance on the hypervisor. [ 842.457396] env[63175]: DEBUG nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 842.458259] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0842ee16-d8eb-41ed-b749-b5b3a422aa59 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.518896] env[63175]: DEBUG nova.scheduler.client.report [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 842.721351] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.721688] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d621577-43a0-402e-9f8a-d52c994f92cc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.728893] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 842.728893] env[63175]: value = "task-1248137" [ 842.728893] env[63175]: _type = "Task" [ 842.728893] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.736832] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.840188] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.866092] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.866453] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Instance network_info: |[{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 842.866961] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:29:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.875712] env[63175]: DEBUG oslo.service.loopingcall [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.875943] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.876181] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab3d49c1-f936-47c2-a3a6-a135b5eca672 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.898679] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.898679] env[63175]: value = "task-1248138" [ 842.898679] env[63175]: _type = "Task" [ 842.898679] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.907101] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248138, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.939418] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.939418] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updated the network info_cache for instance {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.939418] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 842.939418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.967487] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.967761] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.975859] env[63175]: INFO nova.compute.manager [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Took 33.36 seconds to build instance. [ 843.024126] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.024709] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 843.027734] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.603s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.031211] env[63175]: INFO nova.compute.claims [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.239800] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248137, 'name': PowerOffVM_Task, 'duration_secs': 0.268197} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.240166] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.241069] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf953785-5425-4073-be5b-774d885d6aa3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.259483] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3eaeac1-2902-4490-bbde-fc494d68e76d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.290242] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 843.290618] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-809beee0-d8ab-47b8-8c42-7f6ad9383714 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.298027] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 843.298027] env[63175]: value = "task-1248139" [ 843.298027] env[63175]: _type = "Task" [ 843.298027] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.305377] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.346127] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.408824] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248138, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.443061] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.475032] env[63175]: DEBUG nova.compute.utils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.479399] env[63175]: DEBUG oslo_concurrency.lockutils [None req-72034400-3cd6-4726-aba4-78558f87de2f tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.870s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.535162] env[63175]: DEBUG nova.compute.utils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.538975] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 843.539207] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.597128] env[63175]: DEBUG nova.policy [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d2127fd1edc46d0905e20885045f65b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45aaba36fb014a12b34927b2508f0f4d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.808316] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 843.808778] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.809198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.809505] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.809920] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.812021] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-218a6545-f4c1-4f85-a76a-55f68844e315 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.822020] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.822020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.822020] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a5032cf-4b24-474b-b2c1-6e5896f6a718 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.830743] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 843.830743] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52687248-98f4-a9ff-7091-d73dacf2c083" [ 843.830743] env[63175]: _type = "Task" [ 843.830743] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.843774] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52687248-98f4-a9ff-7091-d73dacf2c083, 'name': SearchDatastore_Task, 'duration_secs': 0.010446} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.848116] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.848116] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39e42536-bef1-4685-abe2-14f5ad464c38 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.852881] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 843.852881] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0d2b0-deda-7802-ea44-2c0a90175a4a" [ 843.852881] env[63175]: _type = "Task" [ 843.852881] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.860995] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0d2b0-deda-7802-ea44-2c0a90175a4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.872447] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Successfully created port: bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.911117] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248138, 'name': CreateVM_Task, 'duration_secs': 0.702159} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.911290] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.912140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.912307] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.912652] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.912898] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f0a272-cbf1-49d7-af89-b48805a29cb4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.917536] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 843.917536] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d09418-64d4-109e-ebbf-deb651c95631" [ 843.917536] env[63175]: _type = "Task" [ 843.917536] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.925051] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d09418-64d4-109e-ebbf-deb651c95631, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.981160] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.986227] env[63175]: DEBUG nova.compute.manager [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Received event network-changed-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 843.986227] env[63175]: DEBUG nova.compute.manager [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Refreshing instance network info cache due to event network-changed-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 843.986227] env[63175]: DEBUG oslo_concurrency.lockutils [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] Acquiring lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.986227] env[63175]: DEBUG oslo_concurrency.lockutils [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] Acquired lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.986227] env[63175]: DEBUG nova.network.neutron [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Refreshing network info cache for port a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.046599] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 844.264745] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.264979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.265166] env[63175]: INFO nova.compute.manager [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Shelving [ 844.276845] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd19087-9ac5-4a90-bba7-219ad0bf7284 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.285057] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684a8edd-68c3-4662-a054-7f24fec5830b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.315068] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5afcc9-56dc-4f06-a31b-e392aeb927fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.322475] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8d52e5-dc1c-44d7-9e7e-a2dd1d6e40df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.337490] env[63175]: DEBUG nova.compute.provider_tree [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.346401] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.362716] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0d2b0-deda-7802-ea44-2c0a90175a4a, 'name': SearchDatastore_Task, 'duration_secs': 0.009179} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.362716] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.362904] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. {{(pid=63175) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 844.363187] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d634c54b-222c-4c6c-9954-3f9ab0c7667d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.369574] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 844.369574] env[63175]: value = "task-1248140" [ 844.369574] env[63175]: _type = "Task" [ 844.369574] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.377318] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.427736] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d09418-64d4-109e-ebbf-deb651c95631, 'name': SearchDatastore_Task, 'duration_secs': 0.008454} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.428052] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.428284] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.428512] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.428658] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.428829] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.429096] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9236e99e-5384-47c4-a17b-9c0a230864eb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.436605] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.436793] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.437537] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7d9e848-08c1-4ccb-a96c-f28f18bdf84b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.442408] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 844.442408] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52415198-739b-0613-b422-950c3920e06f" [ 844.442408] env[63175]: _type = "Task" [ 844.442408] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.452891] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52415198-739b-0613-b422-950c3920e06f, 'name': SearchDatastore_Task, 'duration_secs': 0.008195} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.453685] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe543e97-80b9-4ce3-b55c-4a385f48b575 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.458401] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 844.458401] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fb43a3-ba7e-da4a-6162-c5bb7346695a" [ 844.458401] env[63175]: _type = "Task" [ 844.458401] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.465705] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fb43a3-ba7e-da4a-6162-c5bb7346695a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.830317] env[63175]: DEBUG nova.network.neutron [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updated VIF entry in instance network info cache for port a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.830317] env[63175]: DEBUG nova.network.neutron [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.844052] env[63175]: DEBUG nova.scheduler.client.report [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 844.860476] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.881925] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248140, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4953} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.882253] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. [ 844.883078] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036b2f0a-acc9-4355-87af-cd70685e2910 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.909615] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.909936] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3858f1bb-dba3-4769-bf88-43457e65a84e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.930035] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 844.930035] env[63175]: value = "task-1248141" [ 844.930035] env[63175]: _type = "Task" [ 844.930035] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.938363] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248141, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.968654] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fb43a3-ba7e-da4a-6162-c5bb7346695a, 'name': SearchDatastore_Task, 'duration_secs': 0.008186} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.968937] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.969213] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.969475] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9484749a-a9d2-40ba-baab-02c447222772 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.976660] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 844.976660] env[63175]: value = "task-1248142" [ 844.976660] env[63175]: _type = "Task" [ 844.976660] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.985385] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.065736] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.067028] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.067028] env[63175]: INFO nova.compute.manager [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Attaching volume 37c89c77-c578-4a98-b68a-e8d493df963d to /dev/sdb [ 845.069118] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 845.099340] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.099981] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.099981] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.099981] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.100240] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.100270] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.100462] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.100644] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.100815] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.100981] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.101182] env[63175]: DEBUG nova.virt.hardware [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.102062] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa81366-094a-417b-99c8-c8d9888772b3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.110388] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411f9dee-f8e0-4163-8549-17c5acba47a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.115486] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d3e212-436f-486c-8c03-ec724a60c0d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.129849] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397b5439-7da5-487d-954e-d59ea835e5a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.143841] env[63175]: DEBUG nova.virt.block_device [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating existing volume attachment record: cfeb7934-1006-4f76-8f3d-868d7d300427 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 845.280203] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.280582] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae74264a-23c1-4166-8f02-0934c83d1ef1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.291450] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 845.291450] env[63175]: value = "task-1248144" [ 845.291450] env[63175]: _type = "Task" [ 845.291450] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.302975] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.334113] env[63175]: DEBUG oslo_concurrency.lockutils [req-5080a5f5-e9b5-4b88-9c9b-c87d3d648aa6 req-96b47d72-8442-416c-bbf2-29a3dc919f6a service nova] Releasing lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.351617] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.354392] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.355072] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.619s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.355916] env[63175]: DEBUG nova.objects.instance [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lazy-loading 'resources' on Instance uuid cf12aac3-edbb-48eb-b431-70187a4ecda3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.448050] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248141, 'name': ReconfigVM_Task, 'duration_secs': 0.467589} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.448050] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.448050] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1b7837-c299-45fe-9d59-241c9e0e53c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.487832] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3e99ad9-6484-4455-a4e5-7918ba341f86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.501495] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Successfully updated port: bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.508236] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 845.508236] env[63175]: value = "task-1248147" [ 845.508236] env[63175]: _type = "Task" [ 845.508236] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.514253] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248142, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.527399] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248147, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.588258] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.588553] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.801687] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248144, 'name': PowerOffVM_Task, 'duration_secs': 0.318066} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.802064] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.802950] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed73fc4-95c8-4d3e-8fff-89d17e0751e6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.826398] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26a7d97-3730-4533-a8d2-3cbde2dbedbc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.849850] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.861476] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "180e9f26-4fbf-4789-ac58-c462f22588c6" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.861744] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "180e9f26-4fbf-4789-ac58-c462f22588c6" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.991439] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248142, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575978} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.991713] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.991931] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.992196] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79cbbb7b-376a-4897-b433-4f591f884bfd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.998651] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 845.998651] env[63175]: value = "task-1248148" [ 845.998651] env[63175]: _type = "Task" [ 845.998651] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.008301] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.008455] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquired lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.008783] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.012989] env[63175]: DEBUG nova.compute.manager [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Received event network-vif-plugged-bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 846.013215] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Acquiring lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.013546] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.013640] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.013763] env[63175]: DEBUG nova.compute.manager [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] No waiting events found dispatching network-vif-plugged-bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 846.013932] env[63175]: WARNING nova.compute.manager [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Received unexpected event network-vif-plugged-bc362be5-e020-424b-86cd-942f6ceab82f for instance with vm_state building and task_state spawning. [ 846.014133] env[63175]: DEBUG nova.compute.manager [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Received event network-changed-bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 846.014291] env[63175]: DEBUG nova.compute.manager [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Refreshing instance network info cache due to event network-changed-bc362be5-e020-424b-86cd-942f6ceab82f. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 846.014467] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Acquiring lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.018373] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248148, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.033131] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248147, 'name': ReconfigVM_Task, 'duration_secs': 0.182133} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.034121] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.034121] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06e18190-ceea-4a13-9096-e96fdcb27de5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.039979] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 846.039979] env[63175]: value = "task-1248149" [ 846.039979] env[63175]: _type = "Task" [ 846.039979] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.050379] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.091413] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 846.128345] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef67ebe-a26d-4b00-87aa-29f1cdcb67dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.136824] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7affb13a-b1ad-4c16-af69-755aae8e09a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.170776] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20175583-6343-4eaa-964b-1d3312fa141d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.178764] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688b10e5-83cf-4aee-928a-feef8495aa78 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.194024] env[63175]: DEBUG nova.compute.provider_tree [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.339910] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 846.340166] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f679b35f-dcd2-46e1-be60-1769b50ccdd9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.350627] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.351855] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 846.351855] env[63175]: value = "task-1248150" [ 846.351855] env[63175]: _type = "Task" [ 846.351855] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.358944] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248150, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.364426] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "180e9f26-4fbf-4789-ac58-c462f22588c6" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.364923] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 846.509026] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248148, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064679} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.510052] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.510248] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee7dc3d-79f8-4d31-bdf7-74f721e035dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.534904] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.535551] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5369ba2-c339-484d-9ab7-2139e85648f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.551701] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.562389] env[63175]: DEBUG oslo_vmware.api [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248149, 'name': PowerOnVM_Task, 'duration_secs': 0.410177} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.563674] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.565427] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 846.565427] env[63175]: value = "task-1248151" [ 846.565427] env[63175]: _type = "Task" [ 846.565427] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.566363] env[63175]: DEBUG nova.compute.manager [None req-f00e44fa-2a51-4596-9f44-e655e5c0b925 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 846.567146] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e5d584-6159-4021-98f0-a97f6b3652cf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.585350] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248151, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.614482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.690413] env[63175]: DEBUG nova.network.neutron [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Updating instance_info_cache with network_info: [{"id": "bc362be5-e020-424b-86cd-942f6ceab82f", "address": "fa:16:3e:23:03:91", "network": {"id": "91312180-188f-4ac3-9382-de74dd1602d8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1257515736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "45aaba36fb014a12b34927b2508f0f4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc362be5-e0", "ovs_interfaceid": "bc362be5-e020-424b-86cd-942f6ceab82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.700025] env[63175]: DEBUG nova.scheduler.client.report [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 846.849943] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.859829] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248150, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.869247] env[63175]: DEBUG nova.compute.utils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.870645] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 846.870833] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.947021] env[63175]: DEBUG nova.policy [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9398efd4be842f9b6d91c4380984b0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b90bc381bfbf488880d5fa7013bc6e17', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.078622] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248151, 'name': ReconfigVM_Task, 'duration_secs': 0.408327} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.080548] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.080548] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee39a83c-8562-45af-8c43-57cc6c09ab9b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.088240] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 847.088240] env[63175]: value = "task-1248152" [ 847.088240] env[63175]: _type = "Task" [ 847.088240] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.097123] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248152, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.193596] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Releasing lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.194019] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Instance network_info: |[{"id": "bc362be5-e020-424b-86cd-942f6ceab82f", "address": "fa:16:3e:23:03:91", "network": {"id": "91312180-188f-4ac3-9382-de74dd1602d8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1257515736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "45aaba36fb014a12b34927b2508f0f4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc362be5-e0", "ovs_interfaceid": "bc362be5-e020-424b-86cd-942f6ceab82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 847.194529] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Acquired lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.194694] env[63175]: DEBUG nova.network.neutron [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Refreshing network info cache for port bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.196019] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:03:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc362be5-e020-424b-86cd-942f6ceab82f', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.204240] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Creating folder: Project (45aaba36fb014a12b34927b2508f0f4d). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.208294] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.853s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.210648] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db0c5410-ca25-4b40-a058-d6343b4bd643 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.213335] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.467s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.213569] env[63175]: DEBUG nova.objects.instance [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lazy-loading 'resources' on Instance uuid 780989ee-98a1-4e99-9014-45de6921d4bc {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.228356] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Successfully created port: 3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.233666] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Created folder: Project (45aaba36fb014a12b34927b2508f0f4d) in parent group-v268956. [ 847.233800] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Creating folder: Instances. Parent ref: group-v269037. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.234084] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44da71b6-5780-439b-9c30-b0977b2fa8fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.237251] env[63175]: INFO nova.scheduler.client.report [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted allocations for instance cf12aac3-edbb-48eb-b431-70187a4ecda3 [ 847.244475] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Created folder: Instances in parent group-v269037. [ 847.244790] env[63175]: DEBUG oslo.service.loopingcall [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.245216] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 847.245422] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-371f0525-0df6-4ca3-920f-2914f52b8091 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.268923] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.268923] env[63175]: value = "task-1248155" [ 847.268923] env[63175]: _type = "Task" [ 847.268923] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.281331] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248155, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.351325] env[63175]: DEBUG oslo_vmware.api [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248135, 'name': ReconfigVM_Task, 'duration_secs': 5.780902} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.351584] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.351802] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Reconfigured VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 847.363676] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248150, 'name': CreateSnapshot_Task, 'duration_secs': 0.762818} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.364079] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 847.364696] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99f55a1-f1ee-468e-ac6f-a6d8e5cdb899 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.375075] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 847.512575] env[63175]: DEBUG nova.network.neutron [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Updated VIF entry in instance network info cache for port bc362be5-e020-424b-86cd-942f6ceab82f. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.512978] env[63175]: DEBUG nova.network.neutron [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Updating instance_info_cache with network_info: [{"id": "bc362be5-e020-424b-86cd-942f6ceab82f", "address": "fa:16:3e:23:03:91", "network": {"id": "91312180-188f-4ac3-9382-de74dd1602d8", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1257515736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "45aaba36fb014a12b34927b2508f0f4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc362be5-e0", "ovs_interfaceid": "bc362be5-e020-424b-86cd-942f6ceab82f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.604251] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248152, 'name': Rename_Task, 'duration_secs': 0.159978} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.604603] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.604911] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3baaa037-d716-44de-b905-bd04f368931b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.614019] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 847.614019] env[63175]: value = "task-1248157" [ 847.614019] env[63175]: _type = "Task" [ 847.614019] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.622127] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.749171] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4a183d49-1d08-4627-9d6a-65cb5088bb90 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "cf12aac3-edbb-48eb-b431-70187a4ecda3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.557s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.788867] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248155, 'name': CreateVM_Task, 'duration_secs': 0.362617} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.790011] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.791994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.792187] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.792610] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.793066] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81ee634b-25f3-4da1-a897-c46c74ce1e16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.800381] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 847.800381] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5280b89e-6e7d-0342-487a-250cb6b074c5" [ 847.800381] env[63175]: _type = "Task" [ 847.800381] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.809439] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5280b89e-6e7d-0342-487a-250cb6b074c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.891193] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.894412] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7ba6f79b-0fb5-4d53-a935-3a52a0bf2791 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.904053] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 847.904053] env[63175]: value = "task-1248158" [ 847.904053] env[63175]: _type = "Task" [ 847.904053] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.912608] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248158, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.963295] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b4f349-8c81-4cde-b0c0-a5819ac57e0d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.974516] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1ae2ac-b0c8-49c7-9767-33b6c2f32f2a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.011115] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189f1833-96b2-42f3-9638-2833cb195a95 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.015551] env[63175]: DEBUG oslo_concurrency.lockutils [req-a1814946-22bf-4f92-837f-dde3527c8005 req-e9116899-08f6-4b03-ae5c-13db91439dff service nova] Releasing lock "refresh_cache-0bd5141f-0b92-4d27-8d51-023ab1096e78" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.021073] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cd8b19-a634-46a9-a5df-c7ebb41ac6c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.035812] env[63175]: DEBUG nova.compute.provider_tree [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.039118] env[63175]: DEBUG nova.compute.manager [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-vif-deleted-0e60107b-fb39-48fa-8a14-dee31bc1bb00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 848.039343] env[63175]: INFO nova.compute.manager [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Neutron deleted interface 0e60107b-fb39-48fa-8a14-dee31bc1bb00; detaching it from the instance and deleting it from the info cache [ 848.039551] env[63175]: DEBUG nova.network.neutron [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.122959] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248157, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.311587] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5280b89e-6e7d-0342-487a-250cb6b074c5, 'name': SearchDatastore_Task, 'duration_secs': 0.023934} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.311894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.312148] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.312376] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.312521] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.312734] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.313050] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be8eabbe-4a01-4cdb-81c5-798b6ea235de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.320459] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.320632] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.321368] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-945ba3d7-fbfc-4f0f-806d-d798b8d6b0d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.326219] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 848.326219] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f755de-4174-ed2d-bbb8-65a84ccb4be9" [ 848.326219] env[63175]: _type = "Task" [ 848.326219] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.333790] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f755de-4174-ed2d-bbb8-65a84ccb4be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.396105] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 848.413466] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248158, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.421443] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.421574] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.421784] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.421916] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.422072] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.422224] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.422431] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.422617] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.422792] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.422992] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.423131] env[63175]: DEBUG nova.virt.hardware [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.423988] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606b8a2e-e684-4928-b9a3-5f447b4eac43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.431394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b79c0b5-8b1c-4292-8702-9034e8b521e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.543042] env[63175]: DEBUG nova.scheduler.client.report [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 848.547018] env[63175]: DEBUG oslo_concurrency.lockutils [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.547018] env[63175]: DEBUG oslo_concurrency.lockutils [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] Acquired lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.547624] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04a9952-7b70-4081-a6cf-39c35bddcb5e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.565411] env[63175]: DEBUG oslo_concurrency.lockutils [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] Releasing lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.565683] env[63175]: WARNING nova.compute.manager [req-cc0bd149-a199-4be5-83a1-e95bff096b0d req-09c82a31-4ad1-40f1-b95e-7fb53a057052 service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Detach interface failed, port_id=0e60107b-fb39-48fa-8a14-dee31bc1bb00, reason: No device with interface-id 0e60107b-fb39-48fa-8a14-dee31bc1bb00 exists on VM: nova.exception.NotFound: No device with interface-id 0e60107b-fb39-48fa-8a14-dee31bc1bb00 exists on VM [ 848.622869] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248157, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.796837] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.797365] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.797365] env[63175]: DEBUG nova.network.neutron [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.838841] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f755de-4174-ed2d-bbb8-65a84ccb4be9, 'name': SearchDatastore_Task, 'duration_secs': 0.008423} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.841201] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fef2f29d-ab31-4409-a7e7-54f530b4f218 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.849236] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 848.849236] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5234a2b0-7848-32e6-81e4-8cdcf21e8fe5" [ 848.849236] env[63175]: _type = "Task" [ 848.849236] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.858995] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5234a2b0-7848-32e6-81e4-8cdcf21e8fe5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.914244] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248158, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.023482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.023787] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.024000] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.025075] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.025075] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.026850] env[63175]: INFO nova.compute.manager [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Terminating instance [ 849.048268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.053232] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Successfully updated port: 3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.053232] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.018s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.054393] env[63175]: INFO nova.compute.claims [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.091390] env[63175]: INFO nova.scheduler.client.report [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Deleted allocations for instance 780989ee-98a1-4e99-9014-45de6921d4bc [ 849.124865] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248157, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.362028] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5234a2b0-7848-32e6-81e4-8cdcf21e8fe5, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.363058] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.363058] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 0bd5141f-0b92-4d27-8d51-023ab1096e78/0bd5141f-0b92-4d27-8d51-023ab1096e78.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.363058] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33d820be-e44d-41d4-8059-c894d2c78c6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.369994] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 849.369994] env[63175]: value = "task-1248159" [ 849.369994] env[63175]: _type = "Task" [ 849.369994] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.377739] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248159, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.416647] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248158, 'name': CloneVM_Task, 'duration_secs': 1.373884} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.419572] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Created linked-clone VM from snapshot [ 849.420477] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5931a17b-c468-4e12-a05b-4ffa093eb0d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.428352] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Uploading image f69b1429-59f0-4e11-a09d-b5dd7598556d {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 849.455087] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 849.455087] env[63175]: value = "vm-269040" [ 849.455087] env[63175]: _type = "VirtualMachine" [ 849.455087] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 849.455466] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-eb7770bb-c6ca-4216-8e27-b36255105177 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.462708] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lease: (returnval){ [ 849.462708] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d5c6fe-7b70-f62b-9608-b1fbb9c621d3" [ 849.462708] env[63175]: _type = "HttpNfcLease" [ 849.462708] env[63175]: } obtained for exporting VM: (result){ [ 849.462708] env[63175]: value = "vm-269040" [ 849.462708] env[63175]: _type = "VirtualMachine" [ 849.462708] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 849.463164] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the lease: (returnval){ [ 849.463164] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d5c6fe-7b70-f62b-9608-b1fbb9c621d3" [ 849.463164] env[63175]: _type = "HttpNfcLease" [ 849.463164] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.470023] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.470023] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d5c6fe-7b70-f62b-9608-b1fbb9c621d3" [ 849.470023] env[63175]: _type = "HttpNfcLease" [ 849.470023] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 849.482314] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.482645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.533898] env[63175]: DEBUG nova.compute.manager [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 849.533898] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.535481] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ec89d8-13ef-4307-b581-a3d784635edd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.543711] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.544827] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-430aa0f5-5f6e-4fc2-8522-c4748c383300 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.549694] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 849.549694] env[63175]: value = "task-1248161" [ 849.549694] env[63175]: _type = "Task" [ 849.549694] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.558886] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.559051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquired lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.559205] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.562498] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.571256] env[63175]: DEBUG nova.network.neutron [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [{"id": "037fa23d-fc83-4fdc-810b-b97533910bbe", "address": "fa:16:3e:c0:02:c6", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap037fa23d-fc", "ovs_interfaceid": "037fa23d-fc83-4fdc-810b-b97533910bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.599047] env[63175]: DEBUG oslo_concurrency.lockutils [None req-409715fe-2e1e-49e0-8a33-db626c3b6215 tempest-MultipleCreateTestJSON-1550581879 tempest-MultipleCreateTestJSON-1550581879-project-member] Lock "780989ee-98a1-4e99-9014-45de6921d4bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.479s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.623879] env[63175]: DEBUG oslo_vmware.api [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248157, 'name': PowerOnVM_Task, 'duration_secs': 1.589671} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.624131] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.624360] env[63175]: INFO nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Took 8.82 seconds to spawn the instance on the hypervisor. [ 849.624551] env[63175]: DEBUG nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.625395] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9946fb-1800-416f-9656-a1c66253bcb9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.646495] env[63175]: INFO nova.compute.manager [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Rescuing [ 849.646495] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.646716] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.646800] env[63175]: DEBUG nova.network.neutron [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.696393] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 849.696393] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269035', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'name': 'volume-37c89c77-c578-4a98-b68a-e8d493df963d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3', 'attached_at': '', 'detached_at': '', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'serial': '37c89c77-c578-4a98-b68a-e8d493df963d'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 849.697438] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b100b288-bce2-41d8-aab4-1e0d50dd2d7d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.716436] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed42e41-fe44-4e44-879c-8dbfad1df096 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.746029] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] volume-37c89c77-c578-4a98-b68a-e8d493df963d/volume-37c89c77-c578-4a98-b68a-e8d493df963d.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.746879] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acbc5527-76b0-418e-aa26-df8acedce249 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.767023] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 849.767023] env[63175]: value = "task-1248162" [ 849.767023] env[63175]: _type = "Task" [ 849.767023] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.776920] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248162, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.880276] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248159, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.971599] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.971599] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d5c6fe-7b70-f62b-9608-b1fbb9c621d3" [ 849.971599] env[63175]: _type = "HttpNfcLease" [ 849.971599] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 849.971828] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 849.971828] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d5c6fe-7b70-f62b-9608-b1fbb9c621d3" [ 849.971828] env[63175]: _type = "HttpNfcLease" [ 849.971828] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 849.972537] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ab913a-8b7a-48ed-b4b3-fd8c528c2bb3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.980698] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 849.980789] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 850.037430] env[63175]: DEBUG nova.compute.utils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.060497] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248161, 'name': PowerOffVM_Task, 'duration_secs': 0.179064} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.060497] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.060677] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.060798] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06c13243-164d-4d97-bbfe-57a92e478ac8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.074503] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-c445245b-b7d3-49c6-82c5-1e8188c89b68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.097295] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.125742] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e5141799-4eb6-42c5-bd92-edb7a790adbe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.131493] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.131753] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.132075] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleting the datastore file [datastore1] c445245b-b7d3-49c6-82c5-1e8188c89b68 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.134021] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b514ba0-09fb-4568-974b-b69469f1b2e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.145746] env[63175]: INFO nova.compute.manager [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Took 33.88 seconds to build instance. [ 850.151776] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 850.151776] env[63175]: value = "task-1248164" [ 850.151776] env[63175]: _type = "Task" [ 850.151776] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.168060] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.191764] env[63175]: DEBUG nova.compute.manager [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Received event network-vif-plugged-3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 850.195258] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Acquiring lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.195258] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.195662] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.195662] env[63175]: DEBUG nova.compute.manager [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] No waiting events found dispatching network-vif-plugged-3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 850.195662] env[63175]: WARNING nova.compute.manager [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Received unexpected event network-vif-plugged-3d44b610-7af2-4693-a952-96b2a0f3ff98 for instance with vm_state building and task_state spawning. [ 850.196489] env[63175]: DEBUG nova.compute.manager [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Received event network-changed-3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 850.196489] env[63175]: DEBUG nova.compute.manager [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Refreshing instance network info cache due to event network-changed-3d44b610-7af2-4693-a952-96b2a0f3ff98. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 850.196489] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Acquiring lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.279390] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248162, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.294568] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8a33cf-e6e4-4d1b-882f-d8d82c2abf12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.305059] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc455cad-14d1-438c-ab4d-c297024f92df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.341653] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05aecf8e-734f-42e0-9b72-75a97acc0948 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.344771] env[63175]: DEBUG nova.network.neutron [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Updating instance_info_cache with network_info: [{"id": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "address": "fa:16:3e:43:1f:f7", "network": {"id": "544ba266-6c0d-4c1b-88a3-cd340210ff29", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-191168118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b90bc381bfbf488880d5fa7013bc6e17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d44b610-7a", "ovs_interfaceid": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.353258] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef02dee3-1e55-4b92-a62e-be4315c5c4e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.369449] env[63175]: DEBUG nova.compute.provider_tree [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.391032] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248159, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6128} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.391032] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 0bd5141f-0b92-4d27-8d51-023ab1096e78/0bd5141f-0b92-4d27-8d51-023ab1096e78.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.392026] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.392026] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-191fb924-d36e-4b66-b351-48b33b2ed7e6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.398554] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 850.398554] env[63175]: value = "task-1248165" [ 850.398554] env[63175]: _type = "Task" [ 850.398554] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.410390] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248165, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.459660] env[63175]: DEBUG nova.network.neutron [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.540634] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.058s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.583201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f9310e7a-4d6b-4ee2-ba7a-c94c3f2c8743 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-c445245b-b7d3-49c6-82c5-1e8188c89b68-0e60107b-fb39-48fa-8a14-dee31bc1bb00" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.822s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.651034] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c2510710-5c61-479a-854a-b3ac091df589 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.397s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.664480] env[63175]: DEBUG oslo_vmware.api [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392894} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.664787] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.665020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 850.665298] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 850.665538] env[63175]: INFO nova.compute.manager [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Took 1.13 seconds to destroy the instance on the hypervisor. [ 850.665797] env[63175]: DEBUG oslo.service.loopingcall [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.666254] env[63175]: DEBUG nova.compute.manager [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 850.666350] env[63175]: DEBUG nova.network.neutron [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 850.780860] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248162, 'name': ReconfigVM_Task, 'duration_secs': 0.851755} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.781536] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfigured VM instance instance-0000003a to attach disk [datastore2] volume-37c89c77-c578-4a98-b68a-e8d493df963d/volume-37c89c77-c578-4a98-b68a-e8d493df963d.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.791025] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-911a8464-e081-47d1-8609-af980590c812 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.810883] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 850.810883] env[63175]: value = "task-1248166" [ 850.810883] env[63175]: _type = "Task" [ 850.810883] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.818409] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.849966] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Releasing lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.849966] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Instance network_info: |[{"id": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "address": "fa:16:3e:43:1f:f7", "network": {"id": "544ba266-6c0d-4c1b-88a3-cd340210ff29", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-191168118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b90bc381bfbf488880d5fa7013bc6e17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d44b610-7a", "ovs_interfaceid": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 850.849966] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Acquired lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.849966] env[63175]: DEBUG nova.network.neutron [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Refreshing network info cache for port 3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.849966] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:1f:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d44b610-7af2-4693-a952-96b2a0f3ff98', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.863972] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Creating folder: Project (b90bc381bfbf488880d5fa7013bc6e17). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.868644] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63db24ae-c907-4eec-a575-1353d7264b2a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.877818] env[63175]: DEBUG nova.scheduler.client.report [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 850.885481] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Created folder: Project (b90bc381bfbf488880d5fa7013bc6e17) in parent group-v268956. [ 850.885799] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Creating folder: Instances. Parent ref: group-v269041. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.886882] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c7c6435-3fa3-4b19-bfa5-0bfa80968699 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.895552] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Created folder: Instances in parent group-v269041. [ 850.896083] env[63175]: DEBUG oslo.service.loopingcall [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.896186] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.896602] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-926ec9f4-cbfc-4c2c-be42-5d54b1415d07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.930531] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.930531] env[63175]: value = "task-1248169" [ 850.930531] env[63175]: _type = "Task" [ 850.930531] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.930531] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248165, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064941} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.930531] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.934341] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129f7ea5-593d-4b4f-a34a-7ecd8ce6b503 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.943778] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248169, 'name': CreateVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.969717] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 0bd5141f-0b92-4d27-8d51-023ab1096e78/0bd5141f-0b92-4d27-8d51-023ab1096e78.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.970423] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.972990] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2c3e9f0-e6f2-4e17-99f3-b94bc80211c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.994641] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 850.994641] env[63175]: value = "task-1248170" [ 850.994641] env[63175]: _type = "Task" [ 850.994641] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.004541] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248170, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.320885] env[63175]: DEBUG oslo_vmware.api [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248166, 'name': ReconfigVM_Task, 'duration_secs': 0.173798} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.321975] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269035', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'name': 'volume-37c89c77-c578-4a98-b68a-e8d493df963d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3', 'attached_at': '', 'detached_at': '', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'serial': '37c89c77-c578-4a98-b68a-e8d493df963d'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 851.388173] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.388765] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 851.391840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.786s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.394198] env[63175]: INFO nova.compute.claims [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.443605] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248169, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.505263] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248170, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.525287] env[63175]: DEBUG nova.network.neutron [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Updated VIF entry in instance network info cache for port 3d44b610-7af2-4693-a952-96b2a0f3ff98. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.525660] env[63175]: DEBUG nova.network.neutron [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Updating instance_info_cache with network_info: [{"id": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "address": "fa:16:3e:43:1f:f7", "network": {"id": "544ba266-6c0d-4c1b-88a3-cd340210ff29", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-191168118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b90bc381bfbf488880d5fa7013bc6e17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d44b610-7a", "ovs_interfaceid": "3d44b610-7af2-4693-a952-96b2a0f3ff98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.575524] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.575823] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.576087] env[63175]: INFO nova.compute.manager [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Attaching volume 226aa373-bf1a-46ed-b3fe-5461541d261b to /dev/sdb [ 851.625781] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9371517-21fa-4379-8744-934f61edbb36 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.634588] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b09235-79c9-4bd4-b018-88e5ddfc9753 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.648844] env[63175]: DEBUG nova.virt.block_device [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating existing volume attachment record: 884046a0-64f7-4475-baad-5f8a464382b2 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 851.899853] env[63175]: DEBUG nova.compute.utils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.903820] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 851.942083] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248169, 'name': CreateVM_Task, 'duration_secs': 0.543184} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.942336] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.943287] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.943538] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.943994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.944722] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727ab47b-807c-4f46-aeb4-2a1519f6074d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.950728] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 851.950728] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b03134-e07a-196e-dd8a-bd7add471102" [ 851.950728] env[63175]: _type = "Task" [ 851.950728] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.961141] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b03134-e07a-196e-dd8a-bd7add471102, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.009348] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248170, 'name': ReconfigVM_Task, 'duration_secs': 0.529757} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.009348] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 0bd5141f-0b92-4d27-8d51-023ab1096e78/0bd5141f-0b92-4d27-8d51-023ab1096e78.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.010687] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3db517ee-d25b-40c4-b273-b181d482ba75 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.016639] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 852.016639] env[63175]: value = "task-1248173" [ 852.016639] env[63175]: _type = "Task" [ 852.016639] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.024954] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248173, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.028601] env[63175]: DEBUG oslo_concurrency.lockutils [req-644e6f7e-423a-4909-995b-f99fbbdcfbad req-4a99fc0c-cfed-49ff-ada0-e788634fc17c service nova] Releasing lock "refresh_cache-26ab94cc-c5ff-4d69-9d27-f01ca8038b30" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.087262] env[63175]: DEBUG nova.network.neutron [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.287695] env[63175]: DEBUG nova.compute.manager [req-8bf32772-cd6a-4761-adf5-870f1eeb2586 req-e8215b88-0d3b-49e1-ae93-6a079086401a service nova] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Received event network-vif-deleted-037fa23d-fc83-4fdc-810b-b97533910bbe {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 852.389921] env[63175]: DEBUG nova.objects.instance [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.410740] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 852.461519] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b03134-e07a-196e-dd8a-bd7add471102, 'name': SearchDatastore_Task, 'duration_secs': 0.011238} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.465168] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.465480] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.465739] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.465884] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.466082] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.466575] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eccdae05-6e5a-40c5-96b2-c1d6d27bc565 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.478059] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.478059] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.478059] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7401270-95d2-4d22-ba27-b933ec49a73c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.485023] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 852.485023] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f826fc-fb0b-5ece-778c-043155b6f6ae" [ 852.485023] env[63175]: _type = "Task" [ 852.485023] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.494803] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f826fc-fb0b-5ece-778c-043155b6f6ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.527572] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248173, 'name': Rename_Task, 'duration_secs': 0.147194} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.527859] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.528149] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef1a3c66-96e1-4de5-92fe-c89c540436b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.535335] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 852.535335] env[63175]: value = "task-1248175" [ 852.535335] env[63175]: _type = "Task" [ 852.535335] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.543672] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248175, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.546734] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.546734] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f548239-787c-42f7-9702-23debab712b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.552244] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 852.552244] env[63175]: value = "task-1248176" [ 852.552244] env[63175]: _type = "Task" [ 852.552244] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.562833] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248176, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.592543] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.593064] env[63175]: INFO nova.compute.manager [-] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Took 1.93 seconds to deallocate network for instance. [ 852.681440] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd8859f-8a34-4c49-9623-f88be5ac02a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.691309] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071ec1a9-9871-47e5-a0c6-14573b2bd92b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.726683] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d2e137-9d15-428b-bf88-b72439fef6ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.735592] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5809afd-b679-4b01-a9ab-706b3e22a8c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.751629] env[63175]: DEBUG nova.compute.provider_tree [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.895869] env[63175]: DEBUG oslo_concurrency.lockutils [None req-835e8bf5-18d1-4a32-af6b-b6e9942ef710 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.830s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.897059] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.304s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.897059] env[63175]: DEBUG nova.compute.manager [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.898275] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ed8a4f-e429-49dc-81c3-acb0b3d56273 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.906326] env[63175]: DEBUG nova.compute.manager [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63175) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 852.906912] env[63175]: DEBUG nova.objects.instance [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.994426] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f826fc-fb0b-5ece-778c-043155b6f6ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011362} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.995359] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fa1453a-0145-40f4-9760-333223fc0f9b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.001350] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 853.001350] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b4b6d9-cd81-57be-54f2-d3e1533be01a" [ 853.001350] env[63175]: _type = "Task" [ 853.001350] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.010127] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b4b6d9-cd81-57be-54f2-d3e1533be01a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.044915] env[63175]: DEBUG oslo_vmware.api [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248175, 'name': PowerOnVM_Task, 'duration_secs': 0.489867} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.045600] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.045721] env[63175]: INFO nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Took 7.98 seconds to spawn the instance on the hypervisor. [ 853.045856] env[63175]: DEBUG nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 853.046636] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edf02a2-2488-4fb9-a89f-44de3c42dbd6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.062634] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248176, 'name': PowerOffVM_Task, 'duration_secs': 0.195123} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.062991] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.063840] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9146e862-53f4-4190-b45d-9d0267c930ea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.083330] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7436cb1d-159e-47ef-9ee5-81f3ae4bed87 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.102349] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.113039] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.113343] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94929b5e-5b7e-45f5-b3ee-34e7a44262e3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.119833] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 853.119833] env[63175]: value = "task-1248177" [ 853.119833] env[63175]: _type = "Task" [ 853.119833] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.128443] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.254624] env[63175]: DEBUG nova.scheduler.client.report [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 853.420603] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 853.440755] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.441057] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.441224] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.441419] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.441573] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.441726] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.441942] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.442131] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.442309] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.442478] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.442678] env[63175]: DEBUG nova.virt.hardware [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.443532] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e869385-7a77-4861-84f3-f1d3eb56cf03 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.451524] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a9535b-5b3f-4442-9c0e-758fd97a0186 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.467545] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.472999] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Creating folder: Project (bee84536e0d342289dea6e71127fd2c0). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.474699] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b9bdec9-6b4d-4f35-a902-8128b709bdb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.478515] env[63175]: DEBUG nova.compute.manager [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Stashing vm_state: active {{(pid=63175) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 853.485091] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Created folder: Project (bee84536e0d342289dea6e71127fd2c0) in parent group-v268956. [ 853.486074] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Creating folder: Instances. Parent ref: group-v269046. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.486074] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a56b55ae-0394-496e-8460-dd1eae9ec590 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.492970] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Created folder: Instances in parent group-v269046. [ 853.493202] env[63175]: DEBUG oslo.service.loopingcall [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.493381] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.493566] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2719891-b5cd-485b-ab60-82f84adc6c30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.512942] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b4b6d9-cd81-57be-54f2-d3e1533be01a, 'name': SearchDatastore_Task, 'duration_secs': 0.034852} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.514090] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.514351] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 26ab94cc-c5ff-4d69-9d27-f01ca8038b30/26ab94cc-c5ff-4d69-9d27-f01ca8038b30.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.514572] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.514572] env[63175]: value = "task-1248180" [ 853.514572] env[63175]: _type = "Task" [ 853.514572] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.514780] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a160926-b8e0-4ff9-94fd-ec93baa04bb0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.523808] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248180, 'name': CreateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.524999] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 853.524999] env[63175]: value = "task-1248181" [ 853.524999] env[63175]: _type = "Task" [ 853.524999] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.532166] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.567208] env[63175]: INFO nova.compute.manager [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Took 33.31 seconds to build instance. [ 853.631898] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 853.632151] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.632410] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.632557] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.632780] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.633081] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88c04e44-3bf3-436b-9086-531ad714193f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.652926] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.653159] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.654047] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-901630ca-ddcd-41e3-ae6f-d8c6c935f041 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.660309] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 853.660309] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa1b19-7461-f11c-4dc6-d1504c26433e" [ 853.660309] env[63175]: _type = "Task" [ 853.660309] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.668326] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa1b19-7461-f11c-4dc6-d1504c26433e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.759966] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.760689] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 853.764567] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.819s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.764876] env[63175]: DEBUG nova.objects.instance [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lazy-loading 'resources' on Instance uuid 46baf816-eff2-400c-b81c-0d5f3ce8d01e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.914788] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.915192] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18d13718-074b-4317-948e-4d92d696f40f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.925257] env[63175]: DEBUG oslo_vmware.api [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 853.925257] env[63175]: value = "task-1248182" [ 853.925257] env[63175]: _type = "Task" [ 853.925257] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.936956] env[63175]: DEBUG oslo_vmware.api [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248182, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.005490] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.032488] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248180, 'name': CreateVM_Task, 'duration_secs': 0.368112} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.032787] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.033763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.034010] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.034359] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.034640] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e116b5e-8aae-4d1f-a44e-1e9f9180e43d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.040867] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248181, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.042429] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 854.042429] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52db2f67-f188-7b0f-ab9c-36522083a322" [ 854.042429] env[63175]: _type = "Task" [ 854.042429] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.051026] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52db2f67-f188-7b0f-ab9c-36522083a322, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.069626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-177d44dc-2553-4fa0-a032-e119c687319a tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.824s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.175350] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa1b19-7461-f11c-4dc6-d1504c26433e, 'name': SearchDatastore_Task, 'duration_secs': 0.013662} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.176741] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f1a16fb-d8fe-4c22-9a6d-97a4df02a5b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.184152] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 854.184152] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52004469-081f-c326-3add-8bc2487dbee8" [ 854.184152] env[63175]: _type = "Task" [ 854.184152] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.196055] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52004469-081f-c326-3add-8bc2487dbee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.269013] env[63175]: DEBUG nova.compute.utils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.274055] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 854.434554] env[63175]: DEBUG oslo_vmware.api [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248182, 'name': PowerOffVM_Task, 'duration_secs': 0.306471} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.434823] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.435025] env[63175]: DEBUG nova.compute.manager [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 854.435830] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f392d4f8-8de7-4e97-9e98-4b054d3b168e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.473833] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaafae72-d9f0-4159-babc-43c0567faa93 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.483468] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6959b2-ab69-4cc4-9437-d92a47701a0f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.521132] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc0991c-c183-4933-966d-7e11b9733f76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.530710] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cba191-55d6-4d53-b46c-8a569877ecf6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.540156] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580691} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.548288] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 26ab94cc-c5ff-4d69-9d27-f01ca8038b30/26ab94cc-c5ff-4d69-9d27-f01ca8038b30.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.548404] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.548889] env[63175]: DEBUG nova.compute.provider_tree [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.550029] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86fe3789-8004-4caa-bfdd-67570819d3fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.561043] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52db2f67-f188-7b0f-ab9c-36522083a322, 'name': SearchDatastore_Task, 'duration_secs': 0.042719} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.562314] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.562583] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.562920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.563105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.563292] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.563612] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 854.563612] env[63175]: value = "task-1248184" [ 854.563612] env[63175]: _type = "Task" [ 854.563612] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.564459] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96d1d6bb-7583-44b2-90df-7e999f938bdd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.574808] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.576232] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.576434] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.577396] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8c13041-6927-4c41-b4df-47ddd3b1f6f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.582377] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 854.582377] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ae459b-bf7e-614c-7009-bf1990114b57" [ 854.582377] env[63175]: _type = "Task" [ 854.582377] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.590720] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ae459b-bf7e-614c-7009-bf1990114b57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.694517] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52004469-081f-c326-3add-8bc2487dbee8, 'name': SearchDatastore_Task, 'duration_secs': 0.012165} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.694823] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.695117] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. {{(pid=63175) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 854.695393] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-005b610a-a884-43e8-8457-d7348b4752e8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.701576] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 854.701576] env[63175]: value = "task-1248185" [ 854.701576] env[63175]: _type = "Task" [ 854.701576] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.709460] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.772112] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 854.951266] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7f87121b-07f9-4d8b-952b-ab5e22069ba7 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.056107] env[63175]: DEBUG nova.scheduler.client.report [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 855.076290] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063106} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.076606] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.077319] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcbcb78-7606-48e6-8800-9f8c17e6b5fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.098983] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 26ab94cc-c5ff-4d69-9d27-f01ca8038b30/26ab94cc-c5ff-4d69-9d27-f01ca8038b30.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.102265] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-886fe114-9042-43ec-8f8e-ba720c24df14 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.122072] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ae459b-bf7e-614c-7009-bf1990114b57, 'name': SearchDatastore_Task, 'duration_secs': 0.016491} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.123816] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 855.123816] env[63175]: value = "task-1248186" [ 855.123816] env[63175]: _type = "Task" [ 855.123816] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.124076] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55cb6251-cd0c-4352-9ee4-c1eb0c3fbba8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.131539] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 855.131539] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5226d606-9b9e-4a67-9e4b-c2081cdc62ac" [ 855.131539] env[63175]: _type = "Task" [ 855.131539] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.134670] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248186, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.141669] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5226d606-9b9e-4a67-9e4b-c2081cdc62ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.211557] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248185, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.238664] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.239022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.239316] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.239517] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.239705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.242056] env[63175]: INFO nova.compute.manager [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Terminating instance [ 855.561040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.564429] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.121s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.564531] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.564772] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 855.565231] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.951s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.567791] env[63175]: INFO nova.compute.claims [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.571617] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5fabc1-d957-448f-adc5-cc05818c5c65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.583134] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5b918a-adbc-4d38-9ce8-bc800f570e0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.589447] env[63175]: INFO nova.scheduler.client.report [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleted allocations for instance 46baf816-eff2-400c-b81c-0d5f3ce8d01e [ 855.605752] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed89d6-6444-42b9-b787-0ed648f6d860 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.613845] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abd94f4-76d0-47cf-86c2-29031867b3ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.647226] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180929MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 855.647385] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.660137] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248186, 'name': ReconfigVM_Task, 'duration_secs': 0.534692} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.663706] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 26ab94cc-c5ff-4d69-9d27-f01ca8038b30/26ab94cc-c5ff-4d69-9d27-f01ca8038b30.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.664422] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5226d606-9b9e-4a67-9e4b-c2081cdc62ac, 'name': SearchDatastore_Task, 'duration_secs': 0.017012} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.664917] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f828a26e-93ae-46ff-ba2e-27f2b49b84d1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.666624] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.666885] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f37d34d4-b13e-4ce0-b010-3a06feab6324/f37d34d4-b13e-4ce0-b010-3a06feab6324.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.667140] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a656590f-99e8-4a40-9168-4dc61070b296 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.674019] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 855.674019] env[63175]: value = "task-1248188" [ 855.674019] env[63175]: _type = "Task" [ 855.674019] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.675405] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 855.675405] env[63175]: value = "task-1248187" [ 855.675405] env[63175]: _type = "Task" [ 855.675405] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.686596] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248188, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.690339] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248187, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.713107] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.86805} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.714053] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. [ 855.714898] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9248c0ae-8bc0-4493-ba66-21d02495fe86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.742772] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.743462] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7184eac2-1891-443a-9f30-50ac83b85057 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.756673] env[63175]: DEBUG nova.compute.manager [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 855.756880] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.758088] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9366c8bd-e30f-4a7a-8737-e2d58ddea421 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.766325] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.767708] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee6a5e24-cdc7-4b67-8688-0a88e3a0c531 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.769713] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 855.769713] env[63175]: value = "task-1248189" [ 855.769713] env[63175]: _type = "Task" [ 855.769713] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.774821] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 855.774821] env[63175]: value = "task-1248190" [ 855.774821] env[63175]: _type = "Task" [ 855.774821] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.783026] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 855.785195] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248189, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.792224] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.812793] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.813115] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.813324] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.813498] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.813696] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.813955] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.814316] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.814582] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.814817] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.815063] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.815262] env[63175]: DEBUG nova.virt.hardware [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.816313] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21f4d8a-b732-4147-8240-f2f028d84076 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.826686] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fbeefd-d659-4757-a38f-facefbd2a718 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.841264] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.847224] env[63175]: DEBUG oslo.service.loopingcall [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.847557] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.847844] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98928bb8-c13e-41a4-9e1d-be7e3bb1e3e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.864658] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.864658] env[63175]: value = "task-1248191" [ 855.864658] env[63175]: _type = "Task" [ 855.864658] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.873325] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248191, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.108648] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5669264d-1213-4785-869b-c11b88c6dcce tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "46baf816-eff2-400c-b81c-0d5f3ce8d01e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.098s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.191132] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248187, 'name': Rename_Task, 'duration_secs': 0.166595} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.194863] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.195719] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248188, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.196138] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5e81d44-3426-405b-8d83-f2ae90711b49 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.201177] env[63175]: DEBUG nova.objects.instance [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.204641] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 856.204641] env[63175]: value = "task-1248192" [ 856.204641] env[63175]: _type = "Task" [ 856.204641] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.212814] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248192, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.229082] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 856.229993] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269045', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'name': 'volume-226aa373-bf1a-46ed-b3fe-5461541d261b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e4f169-000c-4e9c-8ef5-aa4b4989eb44', 'attached_at': '', 'detached_at': '', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'serial': '226aa373-bf1a-46ed-b3fe-5461541d261b'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 856.230634] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7bf185-be6a-4294-8dd2-4f6fe8867734 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.249882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b85949e-6a21-408e-913d-937b13e1feab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.279160] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] volume-226aa373-bf1a-46ed-b3fe-5461541d261b/volume-226aa373-bf1a-46ed-b3fe-5461541d261b.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.282563] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ea970df-db02-4af4-bd7d-4ad889fb8942 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.307023] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248189, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.308337] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248190, 'name': PowerOffVM_Task, 'duration_secs': 0.225543} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.308628] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 856.308628] env[63175]: value = "task-1248193" [ 856.308628] env[63175]: _type = "Task" [ 856.308628] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.308854] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.309046] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.309341] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7623449a-79d9-4022-96eb-b5ea5c66530b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.319788] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.375018] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248191, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.381232] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.381232] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.381442] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Deleting the datastore file [datastore2] 0bd5141f-0b92-4d27-8d51-023ab1096e78 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.382515] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf2540bb-1e09-48c5-b726-3b8a86521ebb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.388318] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for the task: (returnval){ [ 856.388318] env[63175]: value = "task-1248195" [ 856.388318] env[63175]: _type = "Task" [ 856.388318] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.397195] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248195, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.691662] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248188, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555887} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.694621] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f37d34d4-b13e-4ce0-b010-3a06feab6324/f37d34d4-b13e-4ce0-b010-3a06feab6324.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.694884] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.695421] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21f44f73-9a10-4cc4-acc1-7e96fe4817fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.702148] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 856.702148] env[63175]: value = "task-1248196" [ 856.702148] env[63175]: _type = "Task" [ 856.702148] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.715557] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.715781] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.715915] env[63175]: DEBUG nova.network.neutron [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.716135] env[63175]: DEBUG nova.objects.instance [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'info_cache' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.728455] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248196, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.730598] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248192, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.781805] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248189, 'name': ReconfigVM_Task, 'duration_secs': 0.674474} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.782858] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfigured VM instance instance-00000042 to attach disk [datastore1] a460926e-9637-40aa-bb30-e3890a441e03/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.783769] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b99c0be-1466-4305-a783-e563e65039b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.816514] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f56b4bb-507c-4986-8095-5aa88391066b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.837501] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.839537] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 856.839537] env[63175]: value = "task-1248197" [ 856.839537] env[63175]: _type = "Task" [ 856.839537] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.853582] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248197, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.871221] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1390e77a-066b-4f20-b5f5-cc161994496e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.882839] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5c3585-d77d-43f1-8b17-ff5eeca80d74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.888222] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248191, 'name': CreateVM_Task, 'duration_secs': 0.649767} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.889069] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.889785] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.889981] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.890361] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.894534] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-110a4a36-84cc-4877-9b26-525c759d072c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.926799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e747671-170e-4bfc-a7c4-a51a5d7d3b76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.934907] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 856.934907] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52805906-5010-f4c9-bad9-7dbedb96e70f" [ 856.934907] env[63175]: _type = "Task" [ 856.934907] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.935305] env[63175]: DEBUG oslo_vmware.api [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Task: {'id': task-1248195, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304815} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.936397] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.936397] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.936590] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.936820] env[63175]: INFO nova.compute.manager [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Took 1.18 seconds to destroy the instance on the hypervisor. [ 856.937200] env[63175]: DEBUG oslo.service.loopingcall [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.940811] env[63175]: DEBUG nova.compute.manager [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 856.940927] env[63175]: DEBUG nova.network.neutron [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.947668] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ee1361-eaff-4a19-80b2-ac03217702ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.957650] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52805906-5010-f4c9-bad9-7dbedb96e70f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.972366] env[63175]: DEBUG nova.compute.provider_tree [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.218278] env[63175]: DEBUG oslo_vmware.api [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248192, 'name': PowerOnVM_Task, 'duration_secs': 0.66911} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.221326] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.221562] env[63175]: INFO nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Took 8.83 seconds to spawn the instance on the hypervisor. [ 857.221751] env[63175]: DEBUG nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.222501] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248196, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156899} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.223435] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e40172-9110-44a4-8fc6-653733e709dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.226191] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.226966] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec47fe0-8109-43fa-8599-e12ffebee68c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.230750] env[63175]: DEBUG nova.objects.base [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 857.259214] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] f37d34d4-b13e-4ce0-b010-3a06feab6324/f37d34d4-b13e-4ce0-b010-3a06feab6324.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.260076] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12fb0ef7-34f4-49d9-9b9d-31e9d686e525 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.281486] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 857.281486] env[63175]: value = "task-1248198" [ 857.281486] env[63175]: _type = "Task" [ 857.281486] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.292915] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248198, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.327501] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248193, 'name': ReconfigVM_Task, 'duration_secs': 0.640339} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.327803] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfigured VM instance instance-0000003d to attach disk [datastore2] volume-226aa373-bf1a-46ed-b3fe-5461541d261b/volume-226aa373-bf1a-46ed-b3fe-5461541d261b.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.333529] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4773b1e2-3965-413e-8a7f-438fb49e6998 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.354751] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248197, 'name': ReconfigVM_Task, 'duration_secs': 0.369713} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.356844] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.357248] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 857.357248] env[63175]: value = "task-1248199" [ 857.357248] env[63175]: _type = "Task" [ 857.357248] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.357643] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6befda31-551e-4e9e-85c8-b3ad75c6c3d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.368852] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248199, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.370609] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 857.370609] env[63175]: value = "task-1248200" [ 857.370609] env[63175]: _type = "Task" [ 857.370609] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.380252] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248200, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.447121] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52805906-5010-f4c9-bad9-7dbedb96e70f, 'name': SearchDatastore_Task, 'duration_secs': 0.020384} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.447456] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.447701] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.447945] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.448114] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.448303] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.448590] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-679fb1e3-47d4-4759-b721-dbe83b2bf5b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.458179] env[63175]: DEBUG nova.compute.manager [req-cc0eef01-6fef-4caa-b3bc-d182a3676fef req-d52e62a2-aa29-44e0-864f-3781c0eddbb5 service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Received event network-vif-deleted-bc362be5-e020-424b-86cd-942f6ceab82f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 857.458404] env[63175]: INFO nova.compute.manager [req-cc0eef01-6fef-4caa-b3bc-d182a3676fef req-d52e62a2-aa29-44e0-864f-3781c0eddbb5 service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Neutron deleted interface bc362be5-e020-424b-86cd-942f6ceab82f; detaching it from the instance and deleting it from the info cache [ 857.458598] env[63175]: DEBUG nova.network.neutron [req-cc0eef01-6fef-4caa-b3bc-d182a3676fef req-d52e62a2-aa29-44e0-864f-3781c0eddbb5 service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.466685] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.467083] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.468215] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99452bde-a74e-4c82-8602-d9a71c9e3f17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.476490] env[63175]: DEBUG nova.scheduler.client.report [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 857.480874] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 857.480874] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527b8b99-ebaa-5739-cba1-7731ece26d7f" [ 857.480874] env[63175]: _type = "Task" [ 857.480874] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.490605] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527b8b99-ebaa-5739-cba1-7731ece26d7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.767847] env[63175]: INFO nova.compute.manager [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Took 32.36 seconds to build instance. [ 857.797590] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.873215] env[63175]: DEBUG oslo_vmware.api [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248199, 'name': ReconfigVM_Task, 'duration_secs': 0.182185} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.877046] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269045', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'name': 'volume-226aa373-bf1a-46ed-b3fe-5461541d261b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e4f169-000c-4e9c-8ef5-aa4b4989eb44', 'attached_at': '', 'detached_at': '', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'serial': '226aa373-bf1a-46ed-b3fe-5461541d261b'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 857.885816] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248200, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.933230] env[63175]: DEBUG nova.network.neutron [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.961615] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-961b3071-f96a-4b15-ad09-f73168b993ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.972504] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b27f81-77be-40e9-ae84-095d8ebd64fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.983672] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.984754] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 857.990044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.888s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.990603] env[63175]: DEBUG nova.objects.instance [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'resources' on Instance uuid c445245b-b7d3-49c6-82c5-1e8188c89b68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.002496] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527b8b99-ebaa-5739-cba1-7731ece26d7f, 'name': SearchDatastore_Task, 'duration_secs': 0.017254} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.004458] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81820013-ee33-421d-882b-cca19829a9ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.019230] env[63175]: DEBUG nova.compute.manager [req-cc0eef01-6fef-4caa-b3bc-d182a3676fef req-d52e62a2-aa29-44e0-864f-3781c0eddbb5 service nova] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Detach interface failed, port_id=bc362be5-e020-424b-86cd-942f6ceab82f, reason: Instance 0bd5141f-0b92-4d27-8d51-023ab1096e78 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 858.021213] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 858.021213] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523da62f-288e-845b-fad9-ec7443c39f23" [ 858.021213] env[63175]: _type = "Task" [ 858.021213] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.034831] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523da62f-288e-845b-fad9-ec7443c39f23, 'name': SearchDatastore_Task, 'duration_secs': 0.013354} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.034831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.034831] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.034831] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b0ad29e-e37d-4749-b8f9-ba0166be9139 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.043116] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 858.043116] env[63175]: value = "task-1248201" [ 858.043116] env[63175]: _type = "Task" [ 858.043116] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.053896] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248201, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.228462] env[63175]: DEBUG nova.network.neutron [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.269419] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6c7103d9-a1f3-4728-b89b-aa048ece16db tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.875s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.293644] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248198, 'name': ReconfigVM_Task, 'duration_secs': 0.647162} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.294030] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Reconfigured VM instance instance-00000048 to attach disk [datastore2] f37d34d4-b13e-4ce0-b010-3a06feab6324/f37d34d4-b13e-4ce0-b010-3a06feab6324.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.294751] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abf8b0cf-6af6-4172-97db-86b0672088ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.303037] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 858.303037] env[63175]: value = "task-1248202" [ 858.303037] env[63175]: _type = "Task" [ 858.303037] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.313342] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248202, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.387643] env[63175]: DEBUG oslo_vmware.api [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248200, 'name': PowerOnVM_Task, 'duration_secs': 0.722271} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.387978] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.390783] env[63175]: DEBUG nova.compute.manager [None req-59b02e3a-3e89-49bf-af08-d7e2ef23b05a tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 858.391625] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3b1dff-5f20-4978-b8bd-64e8654f0c86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.437838] env[63175]: INFO nova.compute.manager [-] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Took 1.50 seconds to deallocate network for instance. [ 858.492441] env[63175]: DEBUG nova.compute.utils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.494943] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 858.495280] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.563141] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248201, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.574939] env[63175]: DEBUG nova.policy [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e43774590162406a981cf0d5e20f5a1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83fa8f008a48484195c88ed76ceece0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.733360] env[63175]: DEBUG oslo_concurrency.lockutils [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.771620] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a45d30a-2098-4ad1-bc39-2de139df3d71 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.781617] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b617c0-ca44-4d8c-9764-ac6d0ead0e5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.821648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d08b0c-eeb9-4084-8fec-7b72a9dc84d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.832971] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248202, 'name': Rename_Task, 'duration_secs': 0.232136} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.835620] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f467770-91e2-4522-9d7d-d10b74add704 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.840213] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.840754] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e26d8906-a1c3-4a47-808f-0564f5560fd1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.859353] env[63175]: DEBUG nova.compute.provider_tree [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.862201] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 858.862201] env[63175]: value = "task-1248203" [ 858.862201] env[63175]: _type = "Task" [ 858.862201] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.873390] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.930046] env[63175]: DEBUG nova.objects.instance [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.944351] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.973466] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Successfully created port: e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.999631] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 859.059659] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248201, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.748454} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.059961] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.060968] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.060968] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5449501-539f-4231-9db8-9c1fe7358a27 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.068551] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 859.068551] env[63175]: value = "task-1248204" [ 859.068551] env[63175]: _type = "Task" [ 859.068551] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.077454] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248204, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.332993] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.333256] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.364227] env[63175]: DEBUG nova.scheduler.client.report [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 859.380393] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248203, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.433410] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.433720] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.433984] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.434180] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.434361] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.439184] env[63175]: INFO nova.compute.manager [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Terminating instance [ 859.440625] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8c7a0157-e971-43bf-91c9-010b7c4da1a6 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.865s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.517800] env[63175]: INFO nova.compute.manager [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Unrescuing [ 859.518416] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.518673] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.519057] env[63175]: DEBUG nova.network.neutron [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.580495] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248204, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.216098} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.580911] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.581763] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fc7288-c784-435c-8230-4f5bc7f23426 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.606250] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.606641] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e879ded3-52bc-4a53-b29f-d8f38215f5ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.630779] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 859.630779] env[63175]: value = "task-1248205" [ 859.630779] env[63175]: _type = "Task" [ 859.630779] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.643279] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248205, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.743830] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.744273] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2cefa8a-4e0c-4e75-a2c1-4cb1ce76cdc3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.753283] env[63175]: DEBUG oslo_vmware.api [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 859.753283] env[63175]: value = "task-1248206" [ 859.753283] env[63175]: _type = "Task" [ 859.753283] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.763168] env[63175]: DEBUG oslo_vmware.api [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.835522] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 859.874910] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.877900] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.873s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.886162] env[63175]: DEBUG oslo_vmware.api [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248203, 'name': PowerOnVM_Task, 'duration_secs': 0.820574} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.886540] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.886810] env[63175]: INFO nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Took 6.47 seconds to spawn the instance on the hypervisor. [ 859.887094] env[63175]: DEBUG nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 859.888222] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3261acbf-067e-406d-a151-ba353d373930 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.911729] env[63175]: INFO nova.scheduler.client.report [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted allocations for instance c445245b-b7d3-49c6-82c5-1e8188c89b68 [ 859.946848] env[63175]: DEBUG nova.compute.manager [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 859.947124] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.948395] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981ea2b0-96b3-4e2a-a259-414e1c235448 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.958882] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.959841] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29eb1076-5101-46a5-871b-114d4cce713b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.968280] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 859.968280] env[63175]: value = "task-1248207" [ 859.968280] env[63175]: _type = "Task" [ 859.968280] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.978713] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.012032] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 860.039220] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.039758] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.039973] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.040211] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.040369] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.040529] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.040784] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.040961] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.041182] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.041410] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.041603] env[63175]: DEBUG nova.virt.hardware [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.042548] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415623d5-d3e2-4d98-ac12-06ec5aa940de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.053462] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd18bcc-3cf0-43a5-a46b-466be172c680 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.143307] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248205, 'name': ReconfigVM_Task, 'duration_secs': 0.453786} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.143555] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Reconfigured VM instance instance-00000049 to attach disk [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.144456] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8107c44d-71bc-4217-bb50-fb283f0388c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.153327] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 860.153327] env[63175]: value = "task-1248208" [ 860.153327] env[63175]: _type = "Task" [ 860.153327] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.165575] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248208, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.269649] env[63175]: DEBUG oslo_vmware.api [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.298738] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 860.300028] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f098ded7-e586-4361-9cf0-79a5688041f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.308337] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 860.308495] env[63175]: ERROR oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk due to incomplete transfer. [ 860.308889] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e4151d78-7193-4d87-9869-b16f5d57df74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.316831] env[63175]: DEBUG oslo_vmware.rw_handles [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d18c48-5818-1963-a955-c19cc4144494/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 860.317061] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Uploaded image f69b1429-59f0-4e11-a09d-b5dd7598556d to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 860.320093] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 860.320093] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ec16dec1-bae0-431f-af13-bed58208c72f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.328104] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 860.328104] env[63175]: value = "task-1248209" [ 860.328104] env[63175]: _type = "Task" [ 860.328104] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.343044] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248209, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.359985] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.383994] env[63175]: INFO nova.compute.claims [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.411252] env[63175]: INFO nova.compute.manager [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Took 25.40 seconds to build instance. [ 860.422848] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcb3a4f4-c224-40e1-a540-8e44ddc8fa4f tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "c445245b-b7d3-49c6-82c5-1e8188c89b68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.398s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.480707] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248207, 'name': PowerOffVM_Task, 'duration_secs': 0.315143} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.480808] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.480949] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.481254] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f042c6a8-5f62-40f0-a462-6d0104584e04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.556700] env[63175]: DEBUG nova.network.neutron [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.562267] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.562548] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.562800] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Deleting the datastore file [datastore2] 26ab94cc-c5ff-4d69-9d27-f01ca8038b30 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.563162] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a011a564-9d5c-4505-a5ee-b45e7e760b5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.571276] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for the task: (returnval){ [ 860.571276] env[63175]: value = "task-1248211" [ 860.571276] env[63175]: _type = "Task" [ 860.571276] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.581575] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.667035] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248208, 'name': Rename_Task, 'duration_secs': 0.199257} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.667035] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.667035] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2c13050-3b76-46ce-8d6e-8ecf17cbc44f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.671599] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 860.671599] env[63175]: value = "task-1248212" [ 860.671599] env[63175]: _type = "Task" [ 860.671599] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.680463] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248212, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.768050] env[63175]: DEBUG oslo_vmware.api [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248206, 'name': PowerOnVM_Task, 'duration_secs': 0.584872} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.768357] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.768574] env[63175]: DEBUG nova.compute.manager [None req-05df64ce-49df-4511-816b-f30745871908 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.769490] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cf9b26-e5bf-40aa-8f5a-6752de8aca5e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.788746] env[63175]: DEBUG nova.compute.manager [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Received event network-vif-plugged-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 860.788983] env[63175]: DEBUG oslo_concurrency.lockutils [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] Acquiring lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.789221] env[63175]: DEBUG oslo_concurrency.lockutils [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.789439] env[63175]: DEBUG oslo_concurrency.lockutils [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.789572] env[63175]: DEBUG nova.compute.manager [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] No waiting events found dispatching network-vif-plugged-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 860.789741] env[63175]: WARNING nova.compute.manager [req-281b7210-56c8-4edb-bf20-93cecdea1371 req-4b89462b-f1bd-44c6-a13b-da5bd033fd32 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Received unexpected event network-vif-plugged-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 for instance with vm_state building and task_state spawning. [ 860.840519] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248209, 'name': Destroy_Task} progress is 33%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.890171] env[63175]: INFO nova.compute.resource_tracker [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating resource usage from migration 2a22dd23-126d-4a3c-b064-3e2402ed6681 [ 860.902126] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Successfully updated port: e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.915368] env[63175]: DEBUG oslo_concurrency.lockutils [None req-534d51c3-a8eb-4922-aa43-fd3954e0775a tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.914s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.060247] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.060915] env[63175]: DEBUG nova.objects.instance [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'flavor' on Instance uuid a460926e-9637-40aa-bb30-e3890a441e03 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.083540] env[63175]: DEBUG oslo_vmware.api [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Task: {'id': task-1248211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307767} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.083540] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.083540] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.083697] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.083761] env[63175]: INFO nova.compute.manager [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Took 1.14 seconds to destroy the instance on the hypervisor. [ 861.087766] env[63175]: DEBUG oslo.service.loopingcall [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.087766] env[63175]: DEBUG nova.compute.manager [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 861.087766] env[63175]: DEBUG nova.network.neutron [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.115060] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5f3e8a-9d1b-43c5-ad7e-23233c1848fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.124351] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8f0da7-838b-49d8-a1b8-8f0117508d22 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.161434] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6528add7-5b2f-41a5-9588-82d5a5f8235c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.170551] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b4fd2d-d901-409e-8ef0-5642fb6534aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.188847] env[63175]: DEBUG nova.compute.provider_tree [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.194414] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248212, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.346419] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248209, 'name': Destroy_Task, 'duration_secs': 0.617718} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.346696] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Destroyed the VM [ 861.347349] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 861.347349] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f8554f53-cf4a-4600-8b69-4d73c52aee5c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.360121] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 861.360121] env[63175]: value = "task-1248213" [ 861.360121] env[63175]: _type = "Task" [ 861.360121] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.371800] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248213, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.407728] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.407934] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.408166] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.567134] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a29b9cc-aeb1-4efa-a896-bad02bdc756e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.591363] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.591743] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e2f665b-a3fd-4985-8964-0799ea21a788 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.599501] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 861.599501] env[63175]: value = "task-1248214" [ 861.599501] env[63175]: _type = "Task" [ 861.599501] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.609073] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248214, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.685943] env[63175]: DEBUG oslo_vmware.api [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248212, 'name': PowerOnVM_Task, 'duration_secs': 0.911682} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.686346] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.686480] env[63175]: INFO nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Took 5.90 seconds to spawn the instance on the hypervisor. [ 861.686642] env[63175]: DEBUG nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.687646] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70f4ab5-1fcf-4f9b-90cf-4bb91f45a00b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.697105] env[63175]: DEBUG nova.scheduler.client.report [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 861.872274] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248213, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.914663] env[63175]: DEBUG nova.network.neutron [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.945249] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.064510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.065177] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.111768] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248214, 'name': PowerOffVM_Task, 'duration_secs': 0.394373} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.111990] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 862.118237] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 862.120364] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcc2bfe9-baef-4771-86cc-f0bf4377250a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.140450] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 862.140450] env[63175]: value = "task-1248215" [ 862.140450] env[63175]: _type = "Task" [ 862.140450] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.151126] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.171116] env[63175]: DEBUG nova.network.neutron [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Updating instance_info_cache with network_info: [{"id": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "address": "fa:16:3e:e8:5e:c4", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape68b5b69-c8", "ovs_interfaceid": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.204541] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.327s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.204776] env[63175]: INFO nova.compute.manager [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Migrating [ 862.204985] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.205157] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.208830] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.561s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.210830] env[63175]: INFO nova.compute.manager [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Took 26.63 seconds to build instance. [ 862.360965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.361257] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.372132] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248213, 'name': RemoveSnapshot_Task, 'duration_secs': 0.649279} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.372967] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 862.373266] env[63175]: DEBUG nova.compute.manager [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.374098] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f6b56f-8199-43e6-8bd6-29d179e11101 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.418058] env[63175]: INFO nova.compute.manager [-] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Took 1.33 seconds to deallocate network for instance. [ 862.568257] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 862.651743] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248215, 'name': ReconfigVM_Task, 'duration_secs': 0.284387} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.652131] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 862.652264] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.653041] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1a95f53-00ab-4791-bf6f-f7e36b3fdb93 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.661387] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 862.661387] env[63175]: value = "task-1248216" [ 862.661387] env[63175]: _type = "Task" [ 862.661387] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.670863] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.673485] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.673825] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance network_info: |[{"id": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "address": "fa:16:3e:e8:5e:c4", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape68b5b69-c8", "ovs_interfaceid": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 862.674243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:5e:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '721c6720-3ce0-450e-9951-a894f03acc27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e68b5b69-c8d6-47f5-97a8-96db33cb4eb4', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.681866] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating folder: Project (83fa8f008a48484195c88ed76ceece0e). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.682233] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f248ec08-077e-4c85-a1d4-09a2056c5c1e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.699442] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created folder: Project (83fa8f008a48484195c88ed76ceece0e) in parent group-v268956. [ 862.699629] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating folder: Instances. Parent ref: group-v269050. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.699868] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5bd843c2-11df-47bb-b803-5c0aca236b6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.707713] env[63175]: INFO nova.compute.rpcapi [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 862.708321] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.722813] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c1a9c576-b130-4598-be65-7ad2e62490c7 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.147s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.722813] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created folder: Instances in parent group-v269050. [ 862.723067] env[63175]: DEBUG oslo.service.loopingcall [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.726793] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.731900] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b0abdc9-61f3-4b49-b8be-b52a28f574f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.756732] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.756732] env[63175]: value = "task-1248219" [ 862.756732] env[63175]: _type = "Task" [ 862.756732] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.773321] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248219, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.845976] env[63175]: DEBUG nova.compute.manager [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Received event network-changed-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 862.846709] env[63175]: DEBUG nova.compute.manager [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Refreshing instance network info cache due to event network-changed-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 862.847131] env[63175]: DEBUG oslo_concurrency.lockutils [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] Acquiring lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.847372] env[63175]: DEBUG oslo_concurrency.lockutils [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] Acquired lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.847592] env[63175]: DEBUG nova.network.neutron [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Refreshing network info cache for port e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.866734] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 862.888816] env[63175]: INFO nova.compute.manager [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Shelve offloading [ 862.928742] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.094247] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.172705] env[63175]: DEBUG oslo_vmware.api [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248216, 'name': PowerOnVM_Task, 'duration_secs': 0.449463} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.172896] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.173268] env[63175]: DEBUG nova.compute.manager [None req-e02b2599-433d-4fae-b85d-541dfb10c946 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.174108] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a971d94-90ec-4285-8d8e-60ec08c37e6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.226402] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Applying migration context for instance 4bd91412-c41b-41a1-a648-6b905d826ee3 as it has an incoming, in-progress migration 2a22dd23-126d-4a3c-b064-3e2402ed6681. Migration status is pre-migrating {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 863.226694] env[63175]: INFO nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating resource usage from migration 2a22dd23-126d-4a3c-b064-3e2402ed6681 [ 863.244871] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245168] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245351] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245488] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a460926e-9637-40aa-bb30-e3890a441e03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245607] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245773] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance bfb59d40-c7f2-4f90-9387-8b886d2dbdba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.245936] env[63175]: WARNING nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0bd5141f-0b92-4d27-8d51-023ab1096e78 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 863.246066] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 26ab94cc-c5ff-4d69-9d27-f01ca8038b30 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.246187] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance f37d34d4-b13e-4ce0-b010-3a06feab6324 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.246300] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance f82caee9-bb2f-4c07-b7f0-018ac82c4066 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.246411] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.246523] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Migration 2a22dd23-126d-4a3c-b064-3e2402ed6681 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 863.246693] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 4bd91412-c41b-41a1-a648-6b905d826ee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.248623] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.248863] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.249060] env[63175]: DEBUG nova.network.neutron [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.268287] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248219, 'name': CreateVM_Task, 'duration_secs': 0.401414} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.268743] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.269777] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.269923] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.270473] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.270766] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-170d13a4-62d9-48ea-bc2a-2118a4d13d21 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.275704] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 863.275704] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522890bd-5150-5a00-699e-dae4236ec483" [ 863.275704] env[63175]: _type = "Task" [ 863.275704] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.283922] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522890bd-5150-5a00-699e-dae4236ec483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.385585] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.398645] env[63175]: INFO nova.compute.manager [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Rebuilding instance [ 863.400697] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.401178] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fb464ce-62d4-4cd3-9ac0-c01a23af3585 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.409878] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 863.409878] env[63175]: value = "task-1248220" [ 863.409878] env[63175]: _type = "Task" [ 863.409878] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.426166] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 863.426388] env[63175]: DEBUG nova.compute.manager [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.427411] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dc3143-0aed-449b-96ef-b87e4570a6f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.433798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.433998] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.434164] env[63175]: DEBUG nova.network.neutron [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.441707] env[63175]: DEBUG nova.compute.manager [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.442482] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45fb806-1567-4cf7-b36b-f8d07c173b0d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.548392] env[63175]: DEBUG nova.network.neutron [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Updated VIF entry in instance network info cache for port e68b5b69-c8d6-47f5-97a8-96db33cb4eb4. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.548768] env[63175]: DEBUG nova.network.neutron [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Updating instance_info_cache with network_info: [{"id": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "address": "fa:16:3e:e8:5e:c4", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape68b5b69-c8", "ovs_interfaceid": "e68b5b69-c8d6-47f5-97a8-96db33cb4eb4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.752055] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5d25451b-69aa-4af7-be71-451fa0ac1e39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 863.786844] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522890bd-5150-5a00-699e-dae4236ec483, 'name': SearchDatastore_Task, 'duration_secs': 0.031292} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.787194] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.787435] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.787676] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.787826] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.788018] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.788291] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0eeb54cc-3a90-4bce-ad52-770f093a3d57 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.798088] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.798303] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.799180] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2d643f7-8d8b-49c2-8f1d-88923788b2ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.806135] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 863.806135] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bde13-035b-d912-6873-5e66256cbe96" [ 863.806135] env[63175]: _type = "Task" [ 863.806135] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.815617] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bde13-035b-d912-6873-5e66256cbe96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.050937] env[63175]: DEBUG oslo_concurrency.lockutils [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] Releasing lock "refresh_cache-1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.051539] env[63175]: DEBUG nova.compute.manager [req-44dd603f-45fd-49fd-8d27-29fdda92efeb req-ab0c3652-1ca2-49cb-b903-d963fbe1afb0 service nova] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Received event network-vif-deleted-3d44b610-7af2-4693-a952-96b2a0f3ff98 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 864.265111] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9b72d607-be0f-4caa-b9e2-b25bbbe30f54 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.323478] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bde13-035b-d912-6873-5e66256cbe96, 'name': SearchDatastore_Task, 'duration_secs': 0.012471} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.323731] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f46c6fd-b706-44f8-b370-26a974a79957 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.331092] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 864.331092] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52298fc6-2757-c6cd-2afd-f2f5f3b4a018" [ 864.331092] env[63175]: _type = "Task" [ 864.331092] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.341608] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52298fc6-2757-c6cd-2afd-f2f5f3b4a018, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.438149] env[63175]: DEBUG nova.network.neutron [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.455297] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.455578] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7a038ff-5985-4e05-b52e-900aed22d976 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.462252] env[63175]: DEBUG nova.network.neutron [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updating instance_info_cache with network_info: [{"id": "6f9757fd-9d26-48d8-89e3-3411a8770034", "address": "fa:16:3e:ec:8a:6d", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f9757fd-9d", "ovs_interfaceid": "6f9757fd-9d26-48d8-89e3-3411a8770034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.464705] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 864.464705] env[63175]: value = "task-1248221" [ 864.464705] env[63175]: _type = "Task" [ 864.464705] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.475081] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.767940] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a4304579-f829-433e-a878-1050fac08c2e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.768276] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 864.768436] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 864.842019] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52298fc6-2757-c6cd-2afd-f2f5f3b4a018, 'name': SearchDatastore_Task, 'duration_secs': 0.011943} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.844361] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.844631] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68/1b8b3fdf-642b-40cc-a8ba-07ecd5519a68.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.845079] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae876161-a3b8-4e20-bce9-256656b4ed4c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.852819] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 864.852819] env[63175]: value = "task-1248222" [ 864.852819] env[63175]: _type = "Task" [ 864.852819] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.863440] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248222, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.941392] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.970029] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.980666] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248221, 'name': PowerOffVM_Task, 'duration_secs': 0.212381} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.983809] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.984172] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.987707] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6f4db6-1e9a-4e66-a98b-761e09a1560a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.995829] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 864.995829] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f8aee2e-944d-4f7f-86f8-4e5fd3f26395 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.015811] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f75ada-d6fa-4810-ac59-1d95ccc2aed0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.030110] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2bb5b9-5722-43ba-8eeb-151af472772b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.031841] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.033616] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.033616] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleting the datastore file [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.033616] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c670ac3b-c910-4733-afcd-9a52d7a87392 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.064541] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c1e2bd-1505-4ce4-993e-6569691582f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.067401] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 865.067401] env[63175]: value = "task-1248224" [ 865.067401] env[63175]: _type = "Task" [ 865.067401] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.074923] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cdaffc-0e67-4f0a-a9ca-23d3b9c55041 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.082375] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.096802] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.366353] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248222, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495563} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.366747] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68/1b8b3fdf-642b-40cc-a8ba-07ecd5519a68.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.367008] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.367283] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15903eed-d9b0-48d0-9fea-35e1fb67d6d1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.375428] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 865.375428] env[63175]: value = "task-1248225" [ 865.375428] env[63175]: _type = "Task" [ 865.375428] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.387568] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248225, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.442124] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.443271] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a631b354-62a6-4e61-ad34-f54510d5b60d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.457028] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.457028] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2a33226-aeb8-4683-8d34-d9d76127eda2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.579634] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257254} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.579634] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.579776] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.579951] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.602112] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 865.666560] env[63175]: DEBUG nova.compute.manager [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received event network-vif-unplugged-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 865.666871] env[63175]: DEBUG oslo_concurrency.lockutils [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.667055] env[63175]: DEBUG oslo_concurrency.lockutils [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.667784] env[63175]: DEBUG oslo_concurrency.lockutils [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.667983] env[63175]: DEBUG nova.compute.manager [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] No waiting events found dispatching network-vif-unplugged-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 865.668248] env[63175]: WARNING nova.compute.manager [req-a6342224-a777-4a52-8afa-4b27c13947e3 req-2519d502-3862-48d9-8435-acf666c56b92 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received unexpected event network-vif-unplugged-6f9757fd-9d26-48d8-89e3-3411a8770034 for instance with vm_state shelved and task_state shelving_offloading. [ 865.872276] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.872472] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.872646] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleting the datastore file [datastore1] bfb59d40-c7f2-4f90-9387-8b886d2dbdba {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.872938] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-424f3172-b2e5-4f51-a0a5-8ba14e504f70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.881402] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 865.881402] env[63175]: value = "task-1248227" [ 865.881402] env[63175]: _type = "Task" [ 865.881402] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.884918] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248225, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1299} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.888632] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.889491] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7bc55e-9ffb-49a1-86e6-8b608c36bd6f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.897511] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.915954] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68/1b8b3fdf-642b-40cc-a8ba-07ecd5519a68.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.916282] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a321fbb-4e36-477b-8c8c-408ae42e499d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.941528] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 865.941528] env[63175]: value = "task-1248228" [ 865.941528] env[63175]: _type = "Task" [ 865.941528] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.950472] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.109483] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 866.109773] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.901s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.110019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.166s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.110254] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.114898] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.755s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.116953] env[63175]: INFO nova.compute.claims [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.148319] env[63175]: INFO nova.scheduler.client.report [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Deleted allocations for instance 0bd5141f-0b92-4d27-8d51-023ab1096e78 [ 866.155760] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.156070] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.156270] env[63175]: INFO nova.compute.manager [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Shelving [ 866.394532] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.452108] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248228, 'name': ReconfigVM_Task, 'duration_secs': 0.328585} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.452410] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68/1b8b3fdf-642b-40cc-a8ba-07ecd5519a68.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.453198] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4eb961df-5f4e-4ea7-b473-1722c294c3b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.458357] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceaf64d8-cc67-49ef-aa38-feb4d7bb54d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.464072] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 866.464072] env[63175]: value = "task-1248229" [ 866.464072] env[63175]: _type = "Task" [ 866.464072] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.479126] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 0 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 866.492613] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248229, 'name': Rename_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.621116] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.621380] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.621543] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.621735] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.621886] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.622051] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.622270] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.622435] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.622600] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.622769] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.622993] env[63175]: DEBUG nova.virt.hardware [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.628039] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0cee810-63b5-4807-90ba-8603b9ed58b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.634594] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe30ac7-e354-4897-90d9-8b149121c7b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.650347] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.656371] env[63175]: DEBUG oslo.service.loopingcall [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.659038] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.661031] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65858106-ba3a-4895-be27-26f8dc303167 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.673394] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe9df8ec-b3d4-4709-9258-6ffea0fb5bba tempest-ServerMetadataNegativeTestJSON-2002248185 tempest-ServerMetadataNegativeTestJSON-2002248185-project-member] Lock "0bd5141f-0b92-4d27-8d51-023ab1096e78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.434s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.683596] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.683596] env[63175]: value = "task-1248230" [ 866.683596] env[63175]: _type = "Task" [ 866.683596] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.694377] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248230, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.897493] env[63175]: DEBUG oslo_vmware.api [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.645196} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.898643] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.898643] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.898643] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.919888] env[63175]: INFO nova.scheduler.client.report [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance bfb59d40-c7f2-4f90-9387-8b886d2dbdba [ 866.991500] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.991500] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248229, 'name': Rename_Task, 'duration_secs': 0.175324} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.992080] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-511a3b3e-dad5-430b-a5db-7f3550c0fb8b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.993761] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.994028] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d982a87-014d-41ac-8984-b3bbc048fa88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.002264] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 867.002264] env[63175]: value = "task-1248231" [ 867.002264] env[63175]: _type = "Task" [ 867.002264] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.003987] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 867.003987] env[63175]: value = "task-1248232" [ 867.003987] env[63175]: _type = "Task" [ 867.003987] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.022918] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248232, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.023242] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.182735] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 867.182735] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f7cf65f-77e7-47b0-9d4d-281bd489d4d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.195378] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 867.195378] env[63175]: value = "task-1248233" [ 867.195378] env[63175]: _type = "Task" [ 867.195378] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.198935] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248230, 'name': CreateVM_Task, 'duration_secs': 0.37025} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.202053] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.202696] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.202983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.203393] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.206406] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a0607fa-783e-4b52-9d92-1753c4f60b51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.214501] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248233, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.216040] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 867.216040] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f367be-9cb5-7243-199e-ad8dbface5c0" [ 867.216040] env[63175]: _type = "Task" [ 867.216040] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.228141] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f367be-9cb5-7243-199e-ad8dbface5c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.337598] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.337946] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.374979] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0776cd-c670-4968-9837-e97ba95a2aca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.382542] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab7db09-3531-40cd-85ef-9d431eae3413 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.415344] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af30d657-9ce8-460d-9622-7029bba0ab12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.424717] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.426189] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6328bc-4888-47ce-b012-d71906eb00d1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.441085] env[63175]: DEBUG nova.compute.provider_tree [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.519728] env[63175]: DEBUG oslo_vmware.api [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248232, 'name': PowerOnVM_Task, 'duration_secs': 0.514512} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.522640] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.522937] env[63175]: INFO nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Took 7.51 seconds to spawn the instance on the hypervisor. [ 867.523237] env[63175]: DEBUG nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 867.523515] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248231, 'name': PowerOffVM_Task, 'duration_secs': 0.257288} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.524238] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f239da7-21eb-4d50-b652-49b8d459eb26 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.526915] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.527135] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 17 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 867.695473] env[63175]: DEBUG nova.compute.manager [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Received event network-changed-6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 867.695761] env[63175]: DEBUG nova.compute.manager [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Refreshing instance network info cache due to event network-changed-6f9757fd-9d26-48d8-89e3-3411a8770034. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 867.696091] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] Acquiring lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.696335] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] Acquired lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.696595] env[63175]: DEBUG nova.network.neutron [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Refreshing network info cache for port 6f9757fd-9d26-48d8-89e3-3411a8770034 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.710585] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248233, 'name': PowerOffVM_Task, 'duration_secs': 0.427894} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.710961] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.711792] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320afb10-0197-4ef7-9146-3e8fb3d7aad8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.733893] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a883651d-8548-4adb-b849-f382d9539864 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.739741] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f367be-9cb5-7243-199e-ad8dbface5c0, 'name': SearchDatastore_Task, 'duration_secs': 0.021467} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.740407] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.743460] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.743460] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.743460] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.743460] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.743460] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d744efe-0d66-45c5-956a-cd3bce707a3a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.757289] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.757487] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.758235] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06b2a2eb-09ce-46bf-9b54-97d46de1d9ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.765244] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 867.765244] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290d1ac-c9bf-ddc2-48ce-1a4c52484c90" [ 867.765244] env[63175]: _type = "Task" [ 867.765244] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.777158] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290d1ac-c9bf-ddc2-48ce-1a4c52484c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.843021] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 867.944490] env[63175]: DEBUG nova.scheduler.client.report [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 868.034305] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.034926] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.034926] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.034926] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.035207] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.035350] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.035563] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.035789] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.036247] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.036247] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.036358] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.045374] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c638afdc-146c-47ae-b5c8-f5fed1a8fcfa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.058923] env[63175]: INFO nova.compute.manager [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Took 21.46 seconds to build instance. [ 868.064429] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 868.064429] env[63175]: value = "task-1248234" [ 868.064429] env[63175]: _type = "Task" [ 868.064429] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.073615] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248234, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.249160] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 868.249776] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-affe0e7e-ff3f-496e-9402-799caff1421e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.260401] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 868.260401] env[63175]: value = "task-1248235" [ 868.260401] env[63175]: _type = "Task" [ 868.260401] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.273256] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248235, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.280479] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290d1ac-c9bf-ddc2-48ce-1a4c52484c90, 'name': SearchDatastore_Task, 'duration_secs': 0.023889} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.281268] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a036826-5e50-4f26-bd23-878c07550d83 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.288478] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 868.288478] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b23bc-9239-dea3-24a4-b10937b4b03c" [ 868.288478] env[63175]: _type = "Task" [ 868.288478] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.300881] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b23bc-9239-dea3-24a4-b10937b4b03c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.366812] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.453165] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.453165] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 868.454833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.526s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.455233] env[63175]: DEBUG nova.objects.instance [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lazy-loading 'resources' on Instance uuid 26ab94cc-c5ff-4d69-9d27-f01ca8038b30 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.565430] env[63175]: DEBUG oslo_concurrency.lockutils [None req-210a9d1a-d2a6-45a7-a6a1-0e08b9227256 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.975s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.578183] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.630717] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.635097] env[63175]: DEBUG nova.network.neutron [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updated VIF entry in instance network info cache for port 6f9757fd-9d26-48d8-89e3-3411a8770034. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.635499] env[63175]: DEBUG nova.network.neutron [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updating instance_info_cache with network_info: [{"id": "6f9757fd-9d26-48d8-89e3-3411a8770034", "address": "fa:16:3e:ec:8a:6d", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": null, "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6f9757fd-9d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.724101] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.724449] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.724615] env[63175]: DEBUG nova.compute.manager [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 868.725535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea97718-195a-4e0a-82c7-8054b173c266 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.733398] env[63175]: DEBUG nova.compute.manager [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63175) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 868.734048] env[63175]: DEBUG nova.objects.instance [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'flavor' on Instance uuid 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.771683] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248235, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.800810] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b23bc-9239-dea3-24a4-b10937b4b03c, 'name': SearchDatastore_Task, 'duration_secs': 0.013526} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.801306] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.801611] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.801898] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7fe08cb4-9628-4f5a-a2ae-2a6a1a109a43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.812009] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 868.812009] env[63175]: value = "task-1248236" [ 868.812009] env[63175]: _type = "Task" [ 868.812009] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.821977] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.961737] env[63175]: DEBUG nova.compute.utils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.963437] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 868.963681] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.022912] env[63175]: DEBUG nova.policy [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '782ebbe4b1104fe68f7f5dcd35854a33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975415f67c2645678f0815424128063b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.077496] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.137596] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe6634b4-7ca8-419e-a31d-408670ceebfb req-fde399b3-8894-444f-9cab-505c738baed1 service nova] Releasing lock "refresh_cache-bfb59d40-c7f2-4f90-9387-8b886d2dbdba" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.174300] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c46f1d5-5a17-4609-b70f-40cefbd771dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.182682] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e1fdee-ba8d-45fb-9823-c149ce2dcc5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.216329] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58194f77-f5ae-4716-886d-9951ccf60ed1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.225015] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45e260b-2332-4342-92b1-7dc03219bbfe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.244167] env[63175]: DEBUG nova.compute.provider_tree [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.275229] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248235, 'name': CreateSnapshot_Task, 'duration_secs': 0.887091} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.275604] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 869.276408] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b001760e-e8ed-49a3-b79a-80909ff81656 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.297771] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Successfully created port: cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.323424] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.466753] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 869.576980] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248234, 'name': ReconfigVM_Task, 'duration_secs': 1.28572} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.577357] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 33 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 869.584023] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Successfully created port: dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.746415] env[63175]: DEBUG nova.scheduler.client.report [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 869.754745] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.755328] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9b51b9d-33fd-4006-83d1-1110cd78b2a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.767646] env[63175]: DEBUG oslo_vmware.api [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 869.767646] env[63175]: value = "task-1248237" [ 869.767646] env[63175]: _type = "Task" [ 869.767646] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.788333] env[63175]: DEBUG oslo_vmware.api [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.801421] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 869.802181] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8f534b34-b880-4131-a2ac-c25d44f4ddc3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.816691] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 869.816691] env[63175]: value = "task-1248238" [ 869.816691] env[63175]: _type = "Task" [ 869.816691] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.832905] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248238, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.836799] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248236, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.086066] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.086066] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.086326] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.086326] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.086433] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.086585] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.086793] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.086956] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.087153] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.087326] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.087506] env[63175]: DEBUG nova.virt.hardware [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.093506] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 870.093815] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c50ce0a-04cd-49cb-9ec7-5b769e66e207 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.113864] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 870.113864] env[63175]: value = "task-1248239" [ 870.113864] env[63175]: _type = "Task" [ 870.113864] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.122781] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.256756] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.259062] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.165s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.260626] env[63175]: INFO nova.compute.claims [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.279939] env[63175]: DEBUG oslo_vmware.api [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248237, 'name': PowerOffVM_Task, 'duration_secs': 0.222636} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.280297] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.280545] env[63175]: DEBUG nova.compute.manager [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 870.281261] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e635dee0-08a2-40bb-a0ea-79eaecd6cd31 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.291654] env[63175]: INFO nova.scheduler.client.report [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Deleted allocations for instance 26ab94cc-c5ff-4d69-9d27-f01ca8038b30 [ 870.329569] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248236, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.147551} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.330326] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.330508] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.330741] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2824a1f5-580b-4fb6-bf36-bb2c1a2a47de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.339462] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248238, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.346771] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 870.346771] env[63175]: value = "task-1248240" [ 870.346771] env[63175]: _type = "Task" [ 870.346771] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.356201] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.476590] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 870.502718] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.503061] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.503252] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.503509] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.503720] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.503887] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.504122] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.504296] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.504467] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.504694] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.504801] env[63175]: DEBUG nova.virt.hardware [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.505785] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bed3e72-dd8c-42c8-9a68-af19dee6c1a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.514536] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773355fc-746f-4c7c-b1c6-52004495fb66 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.623549] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248239, 'name': ReconfigVM_Task, 'duration_secs': 0.409179} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.623858] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 870.624692] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dabbad1-bb25-4b2d-9964-34b405e2a9be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.371556] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.372957] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Successfully updated port: cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.377020] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d3917ad7-3e83-45c9-8390-b44aa0b0d16c tempest-ServerGroupTestJSON-1106923568 tempest-ServerGroupTestJSON-1106923568-project-member] Lock "26ab94cc-c5ff-4d69-9d27-f01ca8038b30" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.940s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.377020] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12d2961b-53eb-4f83-ba50-b5fccea319f4 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.651s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.383051] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-360a296c-82ff-479e-8e08-aa2a65f191a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.410789] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27936} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.415625] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.415999] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248238, 'name': CloneVM_Task, 'duration_secs': 1.564639} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.416581] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 871.416581] env[63175]: value = "task-1248241" [ 871.416581] env[63175]: _type = "Task" [ 871.416581] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.416956] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c985c691-07c7-4c01-b35f-7deb77417bbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.419473] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Created linked-clone VM from snapshot [ 871.421055] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb96c97-8c4c-4bed-ab8a-7ee2d5b7f560 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.428599] env[63175]: DEBUG nova.compute.manager [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-vif-plugged-cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 871.428822] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] Acquiring lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.429068] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.429275] env[63175]: DEBUG oslo_concurrency.lockutils [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.429463] env[63175]: DEBUG nova.compute.manager [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] No waiting events found dispatching network-vif-plugged-cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 871.429635] env[63175]: WARNING nova.compute.manager [req-0e49be70-9edb-44de-a85e-21bf99fc7fa0 req-f5038c32-6c77-434e-98cb-93ab4980ebe8 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received unexpected event network-vif-plugged-cba26230-a8fc-4ce1-8729-feb6678721e7 for instance with vm_state building and task_state spawning. [ 871.446933] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.452579] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53d0a3fe-4971-4101-a434-c85be080ce4f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.466844] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Uploading image d8860fd4-87bf-46ae-93cb-6b6f2f8627cd {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 871.468724] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248241, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.476852] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 871.476852] env[63175]: value = "task-1248242" [ 871.476852] env[63175]: _type = "Task" [ 871.476852] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.488031] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248242, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.493657] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 871.493657] env[63175]: value = "vm-269055" [ 871.493657] env[63175]: _type = "VirtualMachine" [ 871.493657] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 871.493949] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3b08e340-5e9c-4b9d-b2fd-45ea70b6d32f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.501542] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lease: (returnval){ [ 871.501542] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163a00-bbb4-ca2a-848c-1abe0778cc8a" [ 871.501542] env[63175]: _type = "HttpNfcLease" [ 871.501542] env[63175]: } obtained for exporting VM: (result){ [ 871.501542] env[63175]: value = "vm-269055" [ 871.501542] env[63175]: _type = "VirtualMachine" [ 871.501542] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 871.501542] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the lease: (returnval){ [ 871.501542] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163a00-bbb4-ca2a-848c-1abe0778cc8a" [ 871.501542] env[63175]: _type = "HttpNfcLease" [ 871.501542] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 871.508597] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 871.508597] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163a00-bbb4-ca2a-848c-1abe0778cc8a" [ 871.508597] env[63175]: _type = "HttpNfcLease" [ 871.508597] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 871.929602] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248241, 'name': ReconfigVM_Task, 'duration_secs': 0.274504} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.930010] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3/4bd91412-c41b-41a1-a648-6b905d826ee3.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.930303] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 50 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.991407] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248242, 'name': ReconfigVM_Task, 'duration_secs': 0.295276} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.991695] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Reconfigured VM instance instance-00000049 to attach disk [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066/f82caee9-bb2f-4c07-b7f0-018ac82c4066.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.992314] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8e4c684-7d9c-40b3-af28-80dc018c407d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.999661] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 871.999661] env[63175]: value = "task-1248244" [ 871.999661] env[63175]: _type = "Task" [ 871.999661] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.013406] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248244, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.013638] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 872.013638] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163a00-bbb4-ca2a-848c-1abe0778cc8a" [ 872.013638] env[63175]: _type = "HttpNfcLease" [ 872.013638] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 872.013928] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 872.013928] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163a00-bbb4-ca2a-848c-1abe0778cc8a" [ 872.013928] env[63175]: _type = "HttpNfcLease" [ 872.013928] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 872.014574] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2339176-abcc-4d10-9d60-08213b76aeb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.023687] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 872.023918] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 872.117631] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a13598dc-d0ec-4d0d-80f6-82c58181112e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.124925] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6a6ced-10ad-47de-a6d4-b7665f18022b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.134858] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67465d13-306f-4246-99cd-3a937545a33c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.169057] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345a9a4f-518f-43f4-881e-56f8c58c94a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.182695] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d974be-2c01-4be3-a571-84d3f52af25d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.197617] env[63175]: DEBUG nova.compute.provider_tree [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.411198] env[63175]: DEBUG nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 872.412448] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e59e60-437c-47cb-96be-03bbbaa99d26 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.439271] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac42de65-acd8-4a06-9f00-e704bc7a06b7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.459795] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631e021c-b931-476f-ac7e-eae5fb108774 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.479536] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 67 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 872.510696] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248244, 'name': Rename_Task, 'duration_secs': 0.139777} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.511235] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.511629] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9eb8a5b0-e358-4b30-9587-b46ec2b668dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.522146] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 872.522146] env[63175]: value = "task-1248245" [ 872.522146] env[63175]: _type = "Task" [ 872.522146] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.531190] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.701712] env[63175]: DEBUG nova.scheduler.client.report [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 872.928239] env[63175]: INFO nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] instance snapshotting [ 872.928470] env[63175]: WARNING nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 872.932535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11334451-c328-4fe7-b3cd-9dc9289d8083 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.957233] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13d2592-d20d-4337-8264-271b5e7e18d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.031966] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248245, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.104856] env[63175]: DEBUG nova.network.neutron [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Port a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf binding to destination host cpu-1 is already ACTIVE {{(pid=63175) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 873.208088] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.949s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.208763] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 873.217439] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.832s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.221712] env[63175]: INFO nova.compute.claims [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.388732] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Successfully updated port: dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.474372] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 873.475444] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4477f6b2-b520-4a52-a673-bd1e12b25bb8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.487707] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-changed-cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 873.488181] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Refreshing instance network info cache due to event network-changed-cba26230-a8fc-4ce1-8729-feb6678721e7. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 873.488648] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Acquiring lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.488975] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Acquired lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.489265] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Refreshing network info cache for port cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.492466] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 873.492466] env[63175]: value = "task-1248246" [ 873.492466] env[63175]: _type = "Task" [ 873.492466] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.508276] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248246, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.536161] env[63175]: DEBUG oslo_vmware.api [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248245, 'name': PowerOnVM_Task, 'duration_secs': 0.638766} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.536493] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.536801] env[63175]: DEBUG nova.compute.manager [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.538015] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d74fe06-b779-42c0-b1ae-55f34476a2d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.726745] env[63175]: DEBUG nova.compute.utils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.733164] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 873.734058] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.776564] env[63175]: DEBUG nova.policy [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '564ff9a1de5f43f4b3b05f3592bfa833', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54aee75c86d543a995a364ed78426ec2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.892419] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.007280] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248246, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.057927] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.080083] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.131336] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.131596] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.131786] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.233660] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 874.242723] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.302511] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Successfully created port: fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.472453] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4be9b52-97c9-4436-97ba-46cac454ed32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.483803] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e79163-688c-4bb3-a4f3-002c651f8405 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.519918] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f182112-4bd2-4252-af64-31d94d6741b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.531702] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc43811-7339-4326-bcdb-58b3c2a32bd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.535887] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248246, 'name': CreateSnapshot_Task, 'duration_secs': 0.723388} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.536260] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 874.537559] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e39921a-2419-4906-8821-71024dc8b232 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.548634] env[63175]: DEBUG nova.compute.provider_tree [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.638684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.639025] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.639317] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.640044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.640044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.641923] env[63175]: INFO nova.compute.manager [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Terminating instance [ 874.749312] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Releasing lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.749657] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-vif-plugged-dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 874.749893] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Acquiring lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.750134] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.750308] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.750501] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] No waiting events found dispatching network-vif-plugged-dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 874.750697] env[63175]: WARNING nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received unexpected event network-vif-plugged-dc3362cd-7410-4a9d-ba7e-35caad1458fb for instance with vm_state building and task_state spawning. [ 874.750930] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-changed-dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 874.751143] env[63175]: DEBUG nova.compute.manager [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Refreshing instance network info cache due to event network-changed-dc3362cd-7410-4a9d-ba7e-35caad1458fb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 874.751336] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Acquiring lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.751483] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Acquired lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.751665] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Refreshing network info cache for port dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.053036] env[63175]: DEBUG nova.scheduler.client.report [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 875.068975] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 875.069903] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3c50b0c4-6d56-4abf-a2ca-489e5851dbae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.079752] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 875.079752] env[63175]: value = "task-1248247" [ 875.079752] env[63175]: _type = "Task" [ 875.079752] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.089470] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248247, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.145798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "refresh_cache-f82caee9-bb2f-4c07-b7f0-018ac82c4066" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.145970] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "refresh_cache-f82caee9-bb2f-4c07-b7f0-018ac82c4066" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.146175] env[63175]: DEBUG nova.network.neutron [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.227400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.227739] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.227908] env[63175]: DEBUG nova.network.neutron [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.254367] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 875.284978] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.285259] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.285424] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.285611] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.285763] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.285910] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.286140] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.286312] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.286487] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.286651] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.286825] env[63175]: DEBUG nova.virt.hardware [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.287708] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04aff797-20d2-408f-aa72-3c37dcd3ec20 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.296308] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0c6284-e3f9-4fd6-95bc-884e53a8f88b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.302582] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.398606] env[63175]: DEBUG nova.network.neutron [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.557809] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.558402] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 875.562021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.137s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.562299] env[63175]: DEBUG nova.objects.instance [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'resources' on Instance uuid bfb59d40-c7f2-4f90-9387-8b886d2dbdba {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.590987] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248247, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.671171] env[63175]: DEBUG nova.network.neutron [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.810346] env[63175]: DEBUG nova.network.neutron [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.863957] env[63175]: DEBUG nova.compute.manager [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Received event network-vif-plugged-fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 875.864363] env[63175]: DEBUG oslo_concurrency.lockutils [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] Acquiring lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.864554] env[63175]: DEBUG oslo_concurrency.lockutils [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.864697] env[63175]: DEBUG oslo_concurrency.lockutils [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.864849] env[63175]: DEBUG nova.compute.manager [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] No waiting events found dispatching network-vif-plugged-fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 875.865042] env[63175]: WARNING nova.compute.manager [req-914a20e6-4094-4e1a-8f6e-022782fb5fea req-b5d1c2ca-79be-4870-a469-fccf7d9fd3df service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Received unexpected event network-vif-plugged-fcbca53b-aabe-41c3-bfdb-0821599b70d8 for instance with vm_state building and task_state spawning. [ 875.902500] env[63175]: DEBUG oslo_concurrency.lockutils [req-1273f4d7-738b-4e2a-a8ab-9a7d30ad121f req-886c1c8c-1ae8-4306-8b49-4c8fa8a58f75 service nova] Releasing lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.902866] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.903161] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.955745] env[63175]: DEBUG nova.network.neutron [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.030493] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Successfully updated port: fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.065396] env[63175]: DEBUG nova.compute.utils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.068047] env[63175]: DEBUG nova.objects.instance [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'numa_topology' on Instance uuid bfb59d40-c7f2-4f90-9387-8b886d2dbdba {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.068608] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 876.068825] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.091419] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248247, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.135183] env[63175]: DEBUG nova.policy [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.316875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "refresh_cache-f82caee9-bb2f-4c07-b7f0-018ac82c4066" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.317482] env[63175]: DEBUG nova.compute.manager [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 876.317721] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.319221] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463a0844-da54-4c30-87eb-16eeafe3ec1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.330121] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.330121] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-556bf019-a264-4839-b4bf-93c1072c16b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.336226] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 876.336226] env[63175]: value = "task-1248248" [ 876.336226] env[63175]: _type = "Task" [ 876.336226] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.346020] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248248, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.453148] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.458628] env[63175]: DEBUG oslo_concurrency.lockutils [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.535819] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.535988] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.536174] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.551893] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Successfully created port: 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.572156] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 876.575338] env[63175]: DEBUG nova.objects.base [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 876.592380] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248247, 'name': CloneVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.791654] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f125bccc-3d6d-4b50-8ed1-c23b5e51771b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.800103] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b413dea0-0acd-47dc-b3d4-871164253d5e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.831661] env[63175]: DEBUG nova.network.neutron [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [{"id": "cba26230-a8fc-4ce1-8729-feb6678721e7", "address": "fa:16:3e:92:70:8b", "network": {"id": "a11a1a05-e5f8-460b-9c02-7d20847bd84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360242438", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcba26230-a8", "ovs_interfaceid": "cba26230-a8fc-4ce1-8729-feb6678721e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dc3362cd-7410-4a9d-ba7e-35caad1458fb", "address": "fa:16:3e:5b:ee:e2", "network": {"id": "ce6cf3f2-d8b9-4cb5-8a6b-c0787472e077", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1420556407", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc3362cd-74", "ovs_interfaceid": "dc3362cd-7410-4a9d-ba7e-35caad1458fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.833482] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab136e2c-76d4-4824-ad9b-a4baa0c1f208 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.842984] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044ec019-8d4d-4608-8fcf-8d232ba098ae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.850876] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248248, 'name': PowerOffVM_Task, 'duration_secs': 0.189059} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.851523] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.851708] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.851956] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-123c1dd0-21c1-4173-a0c2-aa313dfa9dd2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.862125] env[63175]: DEBUG nova.compute.provider_tree [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.889691] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.889944] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.890179] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleting the datastore file [datastore2] f82caee9-bb2f-4c07-b7f0-018ac82c4066 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.890456] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6e27243-d76e-4482-8747-d61e30aad00b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.898642] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 876.898642] env[63175]: value = "task-1248250" [ 876.898642] env[63175]: _type = "Task" [ 876.898642] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.909122] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248250, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.984243] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184a6034-c4c6-4703-b149-62c8cbebc929 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.004297] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058cc0ae-be16-4394-b26b-3733eed5fa9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.013049] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 83 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 877.070499] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.095523] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248247, 'name': CloneVM_Task, 'duration_secs': 1.620748} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.095806] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Created linked-clone VM from snapshot [ 877.096603] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688edacd-94c7-4807-a21e-db30ef06eeb7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.105356] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Uploading image 64e4e8b2-7861-480b-b338-14289e25bd77 {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 877.131959] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 877.131959] env[63175]: value = "vm-269057" [ 877.131959] env[63175]: _type = "VirtualMachine" [ 877.131959] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 877.131959] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f4989bc6-cbf1-442a-8da4-a0552eb68b6a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.142294] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease: (returnval){ [ 877.142294] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5237fe5f-3fd7-8aee-df87-ce06c28ca771" [ 877.142294] env[63175]: _type = "HttpNfcLease" [ 877.142294] env[63175]: } obtained for exporting VM: (result){ [ 877.142294] env[63175]: value = "vm-269057" [ 877.142294] env[63175]: _type = "VirtualMachine" [ 877.142294] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 877.142294] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the lease: (returnval){ [ 877.142294] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5237fe5f-3fd7-8aee-df87-ce06c28ca771" [ 877.142294] env[63175]: _type = "HttpNfcLease" [ 877.142294] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.149829] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.149829] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5237fe5f-3fd7-8aee-df87-ce06c28ca771" [ 877.149829] env[63175]: _type = "HttpNfcLease" [ 877.149829] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.281736] env[63175]: DEBUG nova.network.neutron [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updating instance_info_cache with network_info: [{"id": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "address": "fa:16:3e:4a:fe:51", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbca53b-aa", "ovs_interfaceid": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.336812] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "refresh_cache-5d25451b-69aa-4af7-be71-451fa0ac1e39" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.337200] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance network_info: |[{"id": "cba26230-a8fc-4ce1-8729-feb6678721e7", "address": "fa:16:3e:92:70:8b", "network": {"id": "a11a1a05-e5f8-460b-9c02-7d20847bd84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360242438", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcba26230-a8", "ovs_interfaceid": "cba26230-a8fc-4ce1-8729-feb6678721e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dc3362cd-7410-4a9d-ba7e-35caad1458fb", "address": "fa:16:3e:5b:ee:e2", "network": {"id": "ce6cf3f2-d8b9-4cb5-8a6b-c0787472e077", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1420556407", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc3362cd-74", "ovs_interfaceid": "dc3362cd-7410-4a9d-ba7e-35caad1458fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 877.337778] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:70:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1cbd5e0e-9116-46f1-9748-13a73d2d7e75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cba26230-a8fc-4ce1-8729-feb6678721e7', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:ee:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc3362cd-7410-4a9d-ba7e-35caad1458fb', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.347021] env[63175]: DEBUG oslo.service.loopingcall [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.347267] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.347954] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6ec4b37-e566-43e2-bdc3-91d93311256a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.366904] env[63175]: DEBUG nova.scheduler.client.report [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 877.376839] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.376839] env[63175]: value = "task-1248252" [ 877.376839] env[63175]: _type = "Task" [ 877.376839] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.385357] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248252, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.408733] env[63175]: DEBUG oslo_vmware.api [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248250, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179075} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.409090] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.409296] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.409653] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.409653] env[63175]: INFO nova.compute.manager [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Took 1.09 seconds to destroy the instance on the hypervisor. [ 877.409979] env[63175]: DEBUG oslo.service.loopingcall [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.410166] env[63175]: DEBUG nova.compute.manager [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 877.410260] env[63175]: DEBUG nova.network.neutron [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.431854] env[63175]: DEBUG nova.network.neutron [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.520301] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.520638] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a154da85-eef7-48e7-b6d8-a7b22cf2d65b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.529105] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 877.529105] env[63175]: value = "task-1248253" [ 877.529105] env[63175]: _type = "Task" [ 877.529105] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.537766] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.585976] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 877.608374] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.608691] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.608894] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.609147] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.609346] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.609533] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.609799] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.609979] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.610171] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.610341] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.610514] env[63175]: DEBUG nova.virt.hardware [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.611495] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492822cc-5836-4916-8419-4e0b53786278 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.619943] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92062112-a5f2-47cd-972c-ae2e5ed1ef78 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.652045] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.652045] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5237fe5f-3fd7-8aee-df87-ce06c28ca771" [ 877.652045] env[63175]: _type = "HttpNfcLease" [ 877.652045] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 877.652045] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 877.652045] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5237fe5f-3fd7-8aee-df87-ce06c28ca771" [ 877.652045] env[63175]: _type = "HttpNfcLease" [ 877.652045] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 877.653014] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f961e51-5bc1-47a5-8b8a-0059999f7a86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.662614] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 877.662872] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 877.775275] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c1b174b9-28ff-40b6-b208-a9dc6e5be4ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.784420] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.784734] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Instance network_info: |[{"id": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "address": "fa:16:3e:4a:fe:51", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbca53b-aa", "ovs_interfaceid": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 877.785195] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:fe:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcbca53b-aabe-41c3-bfdb-0821599b70d8', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.793025] env[63175]: DEBUG oslo.service.loopingcall [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.795137] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.795784] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93c225d4-9d8f-417b-bf84-d929baa0db61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.824368] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.824368] env[63175]: value = "task-1248254" [ 877.824368] env[63175]: _type = "Task" [ 877.824368] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.833204] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248254, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.875565] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.313s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.878834] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.512s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.880866] env[63175]: INFO nova.compute.claims [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.896184] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248252, 'name': CreateVM_Task, 'duration_secs': 0.461616} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.896368] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.897273] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.897754] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.898194] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.899548] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5449cd1-26c6-498b-a631-dac1de20d08d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.903031] env[63175]: DEBUG nova.compute.manager [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Received event network-changed-fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 877.903031] env[63175]: DEBUG nova.compute.manager [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Refreshing instance network info cache due to event network-changed-fcbca53b-aabe-41c3-bfdb-0821599b70d8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 877.903161] env[63175]: DEBUG oslo_concurrency.lockutils [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] Acquiring lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.903242] env[63175]: DEBUG oslo_concurrency.lockutils [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] Acquired lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.903418] env[63175]: DEBUG nova.network.neutron [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Refreshing network info cache for port fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.909815] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 877.909815] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526ba5a3-01ed-3885-ac11-88387b53d4c8" [ 877.909815] env[63175]: _type = "Task" [ 877.909815] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.920045] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526ba5a3-01ed-3885-ac11-88387b53d4c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.935045] env[63175]: DEBUG nova.network.neutron [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.042337] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248253, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.190325] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Successfully updated port: 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.336616] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248254, 'name': CreateVM_Task, 'duration_secs': 0.427539} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.336853] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.337506] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.391817] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b57d2099-283c-41d9-bd07-e345f6b90c05 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.127s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.393101] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 9.762s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.393201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.393446] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.393774] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.395564] env[63175]: INFO nova.compute.manager [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Terminating instance [ 878.422417] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526ba5a3-01ed-3885-ac11-88387b53d4c8, 'name': SearchDatastore_Task, 'duration_secs': 0.013344} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.423185] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.423511] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.423855] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.424121] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.424492] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.424784] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.425392] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.425793] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5968b3d-a345-4551-a489-ec2e2a90f32a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.428118] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3e33027-2346-4a5e-bcd7-e05e7204d598 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.434431] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 878.434431] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208b03d-d6cf-db65-aa56-f4f0b6644fb1" [ 878.434431] env[63175]: _type = "Task" [ 878.434431] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.439446] env[63175]: INFO nova.compute.manager [-] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Took 1.03 seconds to deallocate network for instance. [ 878.440865] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.441221] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.444169] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c181db9-9172-42e0-a672-a4b551f1bfcc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.454142] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208b03d-d6cf-db65-aa56-f4f0b6644fb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.461202] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 878.461202] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f31572-593b-1efa-756a-bb16100b873d" [ 878.461202] env[63175]: _type = "Task" [ 878.461202] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.472105] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f31572-593b-1efa-756a-bb16100b873d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.542377] env[63175]: DEBUG oslo_vmware.api [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248253, 'name': PowerOnVM_Task, 'duration_secs': 0.624408} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.543125] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.543125] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-694929ec-5000-4963-aedf-9097a99abaa6 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance '4bd91412-c41b-41a1-a648-6b905d826ee3' progress to 100 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 878.691192] env[63175]: DEBUG nova.network.neutron [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updated VIF entry in instance network info cache for port fcbca53b-aabe-41c3-bfdb-0821599b70d8. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.691773] env[63175]: DEBUG nova.network.neutron [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updating instance_info_cache with network_info: [{"id": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "address": "fa:16:3e:4a:fe:51", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbca53b-aa", "ovs_interfaceid": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.693578] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.693736] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.693877] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.900201] env[63175]: DEBUG nova.compute.manager [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 878.902037] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.902037] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c712bc04-a38c-48ee-91b2-349e7bfa33f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.913893] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7615e8ba-f6b8-4a7d-9e67-a011ece4fb25 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.952549] env[63175]: WARNING nova.virt.vmwareapi.vmops [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bfb59d40-c7f2-4f90-9387-8b886d2dbdba could not be found. [ 878.952914] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.953213] env[63175]: INFO nova.compute.manager [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Took 0.05 seconds to destroy the instance on the hypervisor. [ 878.953549] env[63175]: DEBUG oslo.service.loopingcall [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.961663] env[63175]: DEBUG nova.compute.manager [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 878.961803] env[63175]: DEBUG nova.network.neutron [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.964498] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.964719] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5208b03d-d6cf-db65-aa56-f4f0b6644fb1, 'name': SearchDatastore_Task, 'duration_secs': 0.01601} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.965848] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.966143] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.966319] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.978242] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f31572-593b-1efa-756a-bb16100b873d, 'name': SearchDatastore_Task, 'duration_secs': 0.015618} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.981979] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f862ade3-8fbd-41d4-85a2-269b424cdfc2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.989630] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 878.989630] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527c564b-90a8-d81b-6ab0-b3918006e174" [ 878.989630] env[63175]: _type = "Task" [ 878.989630] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.003426] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527c564b-90a8-d81b-6ab0-b3918006e174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.140860] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ac229a-d35b-40af-968e-92ccaaf917d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.149880] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6557436-55b3-46f0-97ac-5270aa02c8f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.184305] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb75b99-3f2c-4d4f-bbb7-ace89872891d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.192811] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ef8f15-1672-4523-80c0-da8ae66b2483 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.197128] env[63175]: DEBUG oslo_concurrency.lockutils [req-19713bf5-08c7-499a-920c-bffbfa3b07b7 req-af3b15d9-035e-43d7-beb1-1b80f9bc2ec5 service nova] Releasing lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.211167] env[63175]: DEBUG nova.compute.provider_tree [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.230551] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 879.378765] env[63175]: DEBUG nova.network.neutron [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.501776] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527c564b-90a8-d81b-6ab0-b3918006e174, 'name': SearchDatastore_Task, 'duration_secs': 0.025856} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.502117] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.502477] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5d25451b-69aa-4af7-be71-451fa0ac1e39/5d25451b-69aa-4af7-be71-451fa0ac1e39.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.502843] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.503277] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.503555] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c67f34a-e05c-4bb6-9299-2680b3c47b86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.506801] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-185f607e-5d42-46d6-935c-9fae6518e517 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.516888] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 879.516888] env[63175]: value = "task-1248255" [ 879.516888] env[63175]: _type = "Task" [ 879.516888] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.522745] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.523118] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.527130] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f61b6932-104d-49ff-bb78-4ad531fcf191 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.529635] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.534432] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 879.534432] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5278e293-9faa-e918-c8bb-d24a4eda1e10" [ 879.534432] env[63175]: _type = "Task" [ 879.534432] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.544801] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5278e293-9faa-e918-c8bb-d24a4eda1e10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.715453] env[63175]: DEBUG nova.scheduler.client.report [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 879.739719] env[63175]: DEBUG nova.network.neutron [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.881743] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.882084] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Instance network_info: |[{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 879.882875] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:3e:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cd631a8-97d6-4ff8-87c6-ab78b6d39c51', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.891037] env[63175]: DEBUG oslo.service.loopingcall [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.891338] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.891763] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11f1a1ab-34f8-4e1f-a068-797351328479 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.913464] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.913464] env[63175]: value = "task-1248256" [ 879.913464] env[63175]: _type = "Task" [ 879.913464] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.924372] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248256, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.012551] env[63175]: DEBUG nova.compute.manager [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-plugged-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 880.012551] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.012551] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.012551] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.012551] env[63175]: DEBUG nova.compute.manager [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] No waiting events found dispatching network-vif-plugged-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 880.012551] env[63175]: WARNING nova.compute.manager [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received unexpected event network-vif-plugged-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 for instance with vm_state building and task_state spawning. [ 880.012551] env[63175]: DEBUG nova.compute.manager [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-changed-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 880.013098] env[63175]: DEBUG nova.compute.manager [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing instance network info cache due to event network-changed-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 880.013507] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.014301] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.015121] env[63175]: DEBUG nova.network.neutron [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing network info cache for port 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.030791] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248255, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.049362] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5278e293-9faa-e918-c8bb-d24a4eda1e10, 'name': SearchDatastore_Task, 'duration_secs': 0.013241} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.049362] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2b0f87d-ab71-41a2-8ae3-a0ee8d5ba28d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.054934] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 880.054934] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52959864-5916-4f47-4bbd-f73a9fc0ade9" [ 880.054934] env[63175]: _type = "Task" [ 880.054934] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.077375] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52959864-5916-4f47-4bbd-f73a9fc0ade9, 'name': SearchDatastore_Task, 'duration_secs': 0.01426} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.082607] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.082607] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9b72d607-be0f-4caa-b9e2-b25bbbe30f54/9b72d607-be0f-4caa-b9e2-b25bbbe30f54.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.082804] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e82dcb78-3d54-4b92-92d6-533c739420e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.092037] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 880.092037] env[63175]: value = "task-1248257" [ 880.092037] env[63175]: _type = "Task" [ 880.092037] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.105445] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248257, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.151788] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 880.152619] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0c4f2f-996c-4fd1-a8f7-bd68ab3654e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.160301] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 880.160479] env[63175]: ERROR oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk due to incomplete transfer. [ 880.160731] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2692a845-2776-4691-aa4a-9e7e24bbe04f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.170831] env[63175]: DEBUG oslo_vmware.rw_handles [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523dda04-75cf-4cb3-5ad9-cd9b94cc5b4a/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 880.171071] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Uploaded image d8860fd4-87bf-46ae-93cb-6b6f2f8627cd to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 880.173328] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 880.173609] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d8f2ee95-f16d-42cd-9480-a11ce8280a79 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.181735] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 880.181735] env[63175]: value = "task-1248258" [ 880.181735] env[63175]: _type = "Task" [ 880.181735] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.193278] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248258, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.221617] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.222247] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 880.225339] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.168s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.225544] env[63175]: DEBUG nova.objects.instance [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 880.241993] env[63175]: INFO nova.compute.manager [-] [instance: bfb59d40-c7f2-4f90-9387-8b886d2dbdba] Took 1.28 seconds to deallocate network for instance. [ 880.425676] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248256, 'name': CreateVM_Task, 'duration_secs': 0.456282} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.425897] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.426599] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.426778] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.427134] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.427411] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e03eff86-b935-4022-97ec-e081edc2b041 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.434860] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 880.434860] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52406c55-ae38-a181-1ab9-87ec5a87c6fd" [ 880.434860] env[63175]: _type = "Task" [ 880.434860] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.444614] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52406c55-ae38-a181-1ab9-87ec5a87c6fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.531742] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554262} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.532210] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5d25451b-69aa-4af7-be71-451fa0ac1e39/5d25451b-69aa-4af7-be71-451fa0ac1e39.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.532337] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.532622] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12320672-df7d-487f-a0ac-1742cf367a9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.542317] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 880.542317] env[63175]: value = "task-1248259" [ 880.542317] env[63175]: _type = "Task" [ 880.542317] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.554420] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248259, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.602722] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248257, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.694185] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248258, 'name': Destroy_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.737163] env[63175]: DEBUG nova.compute.utils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.743974] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 880.744181] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.748361] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 880.805711] env[63175]: DEBUG nova.policy [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1502badb089644d4a06e6bec34e45a8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822ae9f380c84b3c914770f1232975b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.879656] env[63175]: DEBUG nova.network.neutron [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updated VIF entry in instance network info cache for port 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.880047] env[63175]: DEBUG nova.network.neutron [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.947528] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52406c55-ae38-a181-1ab9-87ec5a87c6fd, 'name': SearchDatastore_Task, 'duration_secs': 0.024608} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.947895] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.948146] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.948398] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.948553] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.948738] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.949031] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f6725c1-26a5-4a19-94a7-8027b6faf9bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.962647] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.962856] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.963693] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-766acbb2-3123-4dbb-865e-aab40624e5d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.970285] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 880.970285] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5263de5b-d05c-5abd-132d-0fc284fdf975" [ 880.970285] env[63175]: _type = "Task" [ 880.970285] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.979441] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5263de5b-d05c-5abd-132d-0fc284fdf975, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.053543] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248259, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.161785} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.053834] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.054738] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede77688-df5f-44d6-89f7-b8f6bafa05d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.089421] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 5d25451b-69aa-4af7-be71-451fa0ac1e39/5d25451b-69aa-4af7-be71-451fa0ac1e39.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.089776] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ab2435e-f288-486c-b847-0c09923a8fa8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.114823] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248257, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579077} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.115921] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9b72d607-be0f-4caa-b9e2-b25bbbe30f54/9b72d607-be0f-4caa-b9e2-b25bbbe30f54.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.116179] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.116641] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 881.116641] env[63175]: value = "task-1248260" [ 881.116641] env[63175]: _type = "Task" [ 881.116641] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.116861] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-890aeb90-a023-4c98-a8ff-e12450332074 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.138458] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248260, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.142039] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 881.142039] env[63175]: value = "task-1248261" [ 881.142039] env[63175]: _type = "Task" [ 881.142039] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.151865] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.180320] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Successfully created port: a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.193430] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248258, 'name': Destroy_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.239550] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aa91b4f9-23df-4f72-8ab0-4f9912a7f660 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.240824] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.276s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.240987] env[63175]: DEBUG nova.objects.instance [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lazy-loading 'resources' on Instance uuid f82caee9-bb2f-4c07-b7f0-018ac82c4066 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.277721] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0ef99d96-0b6b-43de-befa-4f8556d3cf4a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "bfb59d40-c7f2-4f90-9387-8b886d2dbdba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.883s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.290948] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.291963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.291963] env[63175]: DEBUG nova.compute.manager [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Going to confirm migration 1 {{(pid=63175) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 881.384905] env[63175]: DEBUG oslo_concurrency.lockutils [req-c43bfa47-8a3f-47a8-be92-7e796a37f5a6 req-739956b9-62bd-4db3-bafc-068038bd92d1 service nova] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.481546] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5263de5b-d05c-5abd-132d-0fc284fdf975, 'name': SearchDatastore_Task, 'duration_secs': 0.017732} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.482362] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c79148-8eaf-4983-8ccd-2ffa965eb10e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.488987] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 881.488987] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5269b22a-88f1-92d1-825a-63f5514edd68" [ 881.488987] env[63175]: _type = "Task" [ 881.488987] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.503025] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5269b22a-88f1-92d1-825a-63f5514edd68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.637934] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248260, 'name': ReconfigVM_Task, 'duration_secs': 0.433312} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.638241] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 5d25451b-69aa-4af7-be71-451fa0ac1e39/5d25451b-69aa-4af7-be71-451fa0ac1e39.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.638954] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a08ad9f-caa9-4fb7-a1a3-806e2b0c370e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.650253] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.509147} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.651693] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.652083] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 881.652083] env[63175]: value = "task-1248262" [ 881.652083] env[63175]: _type = "Task" [ 881.652083] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.652795] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a00bfa-a39f-40a1-9892-9335519cfe9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.664384] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248262, 'name': Rename_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.682540] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 9b72d607-be0f-4caa-b9e2-b25bbbe30f54/9b72d607-be0f-4caa-b9e2-b25bbbe30f54.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.682920] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-056bcd8e-170f-41e3-a781-894a2971e955 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.708245] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248258, 'name': Destroy_Task, 'duration_secs': 1.422866} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.710390] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Destroyed the VM [ 881.710678] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 881.711205] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 881.711205] env[63175]: value = "task-1248263" [ 881.711205] env[63175]: _type = "Task" [ 881.711205] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.711455] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c2b9e8d6-d375-44eb-a41b-010e78302dba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.723301] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248263, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.725025] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 881.725025] env[63175]: value = "task-1248264" [ 881.725025] env[63175]: _type = "Task" [ 881.725025] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.736093] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248264, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.760438] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 881.784355] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.784739] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.784894] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.785064] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.785250] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.785412] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.785653] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.785823] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.786024] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.786824] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.786824] env[63175]: DEBUG nova.virt.hardware [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.787329] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138e8591-1e60-42cf-bd08-22fae67bcfe7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.802671] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e222284-f212-4426-81f7-fdc771b077ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.916893] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.917158] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.917386] env[63175]: DEBUG nova.network.neutron [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.917645] env[63175]: DEBUG nova.objects.instance [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'info_cache' on Instance uuid 4bd91412-c41b-41a1-a648-6b905d826ee3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.964508] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1108ee-6219-4992-a956-dff2beacce98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.973601] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae28e83-ce7b-42a0-905e-b7b2e828de69 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.010999] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf3bb9-a15c-4f58-985d-be445cd1e9f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.022797] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98844900-a9d1-431b-b87c-05db9407b445 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.026752] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5269b22a-88f1-92d1-825a-63f5514edd68, 'name': SearchDatastore_Task, 'duration_secs': 0.024369} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.027057] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.027321] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a4304579-f829-433e-a878-1050fac08c2e/a4304579-f829-433e-a878-1050fac08c2e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.027941] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4c7df63-0775-4e74-a221-33858557507b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.037920] env[63175]: DEBUG nova.compute.provider_tree [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.046628] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 882.046628] env[63175]: value = "task-1248265" [ 882.046628] env[63175]: _type = "Task" [ 882.046628] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.057999] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.166359] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248262, 'name': Rename_Task, 'duration_secs': 0.187056} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.166570] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.166865] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1294c17c-76f1-4dbd-95be-a4c07c2b2f17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.174752] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 882.174752] env[63175]: value = "task-1248266" [ 882.174752] env[63175]: _type = "Task" [ 882.174752] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.183321] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.223730] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248263, 'name': ReconfigVM_Task, 'duration_secs': 0.385411} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.224068] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 9b72d607-be0f-4caa-b9e2-b25bbbe30f54/9b72d607-be0f-4caa-b9e2-b25bbbe30f54.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.224722] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c66aea24-9114-4daf-9084-17fbc828e15c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.235662] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248264, 'name': RemoveSnapshot_Task, 'duration_secs': 0.388283} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.237116] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 882.237431] env[63175]: DEBUG nova.compute.manager [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.237856] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 882.237856] env[63175]: value = "task-1248267" [ 882.237856] env[63175]: _type = "Task" [ 882.237856] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.238515] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f491907-ffcd-4149-8782-b5ce6baa3c23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.250915] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248267, 'name': Rename_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.279140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.279600] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.541665] env[63175]: DEBUG nova.scheduler.client.report [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 882.559171] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248265, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.692297] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248266, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.703431] env[63175]: DEBUG nova.compute.manager [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-vif-plugged-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 882.703676] env[63175]: DEBUG oslo_concurrency.lockutils [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.703891] env[63175]: DEBUG oslo_concurrency.lockutils [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] Lock "0645fffe-9910-43af-af40-126592aefaff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.704116] env[63175]: DEBUG oslo_concurrency.lockutils [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] Lock "0645fffe-9910-43af-af40-126592aefaff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.704259] env[63175]: DEBUG nova.compute.manager [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] No waiting events found dispatching network-vif-plugged-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 882.704438] env[63175]: WARNING nova.compute.manager [req-c23004d5-767a-436d-9445-ac47cd2d1ee5 req-3ec805de-4ac0-4f79-89b4-5a491db6efdc service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received unexpected event network-vif-plugged-a16e4e4f-df78-450a-bf4d-d43daac38791 for instance with vm_state building and task_state spawning. [ 882.751963] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248267, 'name': Rename_Task, 'duration_secs': 0.201463} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.752451] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.752857] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8c67acb-4cae-481c-a558-b5e47cc87e1f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.759148] env[63175]: INFO nova.compute.manager [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Shelve offloading [ 882.764377] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 882.764377] env[63175]: value = "task-1248268" [ 882.764377] env[63175]: _type = "Task" [ 882.764377] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.774342] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.781636] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 882.882179] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Successfully updated port: a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.052356] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.067375] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248265, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.617036} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.067560] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a4304579-f829-433e-a878-1050fac08c2e/a4304579-f829-433e-a878-1050fac08c2e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.067673] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.067869] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5a1b8f7-ce6a-49a1-ba27-72b820daa595 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.078113] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 883.078113] env[63175]: value = "task-1248269" [ 883.078113] env[63175]: _type = "Task" [ 883.078113] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.079755] env[63175]: INFO nova.scheduler.client.report [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleted allocations for instance f82caee9-bb2f-4c07-b7f0-018ac82c4066 [ 883.095606] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248269, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.190572] env[63175]: DEBUG oslo_vmware.api [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248266, 'name': PowerOnVM_Task, 'duration_secs': 0.618576} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.190934] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.191150] env[63175]: INFO nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Took 12.71 seconds to spawn the instance on the hypervisor. [ 883.191370] env[63175]: DEBUG nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.192297] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b73ef82-ed1f-46ea-8851-2ac39b3f16da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.223029] env[63175]: DEBUG nova.network.neutron [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [{"id": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "address": "fa:16:3e:f4:29:ce", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30697ac-b3", "ovs_interfaceid": "a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.263675] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.263984] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a05bcce-87f5-4e29-8dd9-75122be77435 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.274683] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.276097] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 883.276097] env[63175]: value = "task-1248270" [ 883.276097] env[63175]: _type = "Task" [ 883.276097] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.287251] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 883.287645] env[63175]: DEBUG nova.compute.manager [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.290068] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb735dec-265e-4daa-ace7-202c3157467e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.298268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.298499] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.298676] env[63175]: DEBUG nova.network.neutron [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.304199] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.304505] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.305996] env[63175]: INFO nova.compute.claims [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.386822] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.386822] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.386822] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.592359] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248269, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089653} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.592796] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c3983f4b-e49a-43cb-8b8f-5b1552513da3 tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f82caee9-bb2f-4c07-b7f0-018ac82c4066" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.954s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.593784] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.594814] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c34ca6b-f88b-4cc1-ba78-26ce57a810a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.617832] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] a4304579-f829-433e-a878-1050fac08c2e/a4304579-f829-433e-a878-1050fac08c2e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.618804] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d78f8682-ccc7-461d-a726-21b854b2f6f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.639071] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 883.639071] env[63175]: value = "task-1248271" [ 883.639071] env[63175]: _type = "Task" [ 883.639071] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.647666] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.712916] env[63175]: INFO nova.compute.manager [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Took 23.37 seconds to build instance. [ 883.725140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-4bd91412-c41b-41a1-a648-6b905d826ee3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.725418] env[63175]: DEBUG nova.objects.instance [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'migration_context' on Instance uuid 4bd91412-c41b-41a1-a648-6b905d826ee3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.777089] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248268, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.924172] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.077713] env[63175]: DEBUG nova.network.neutron [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.103847] env[63175]: DEBUG nova.network.neutron [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.150096] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248271, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.214680] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f3443f4-9ffc-4927-b943-e38edb1bfd28 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.881s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.227967] env[63175]: DEBUG nova.objects.base [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Object Instance<4bd91412-c41b-41a1-a648-6b905d826ee3> lazy-loaded attributes: info_cache,migration_context {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 884.229270] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8ef79f-dd82-4953-aaed-572adc638c20 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.249727] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9c7e153-7804-4cb1-ac2c-a45421511caf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.257999] env[63175]: DEBUG oslo_vmware.api [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 884.257999] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52485973-5351-5f03-1206-8fcf40a6fd24" [ 884.257999] env[63175]: _type = "Task" [ 884.257999] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.267900] env[63175]: DEBUG oslo_vmware.api [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52485973-5351-5f03-1206-8fcf40a6fd24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.276507] env[63175]: DEBUG oslo_vmware.api [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248268, 'name': PowerOnVM_Task, 'duration_secs': 1.185775} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.276778] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.276990] env[63175]: INFO nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Took 9.02 seconds to spawn the instance on the hypervisor. [ 884.277196] env[63175]: DEBUG nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 884.278141] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1429062c-1fdf-40e7-a543-76cc9ae3ca91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.507491] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81325a40-3c09-493a-b7ab-b453e4aa8300 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.517227] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66d7f2a-8fcc-4418-91c9-3119370270f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.550234] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb947be-50d1-4697-b387-813efdd6f8fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.553077] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.553328] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.553535] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "f37d34d4-b13e-4ce0-b010-3a06feab6324-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.553717] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.553884] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.556577] env[63175]: INFO nova.compute.manager [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Terminating instance [ 884.562062] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90ff015-5d9c-4db8-bc8f-70175046cd61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.576629] env[63175]: DEBUG nova.compute.provider_tree [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.581013] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.581013] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Instance network_info: |[{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 884.581231] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:a2:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0467beaa-08c6-44d6-b8a2-e9c609c21ff4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a16e4e4f-df78-450a-bf4d-d43daac38791', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.588556] env[63175]: DEBUG oslo.service.loopingcall [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.589250] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.589477] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75a979d8-31d6-4e9a-9ac8-99b21007bf88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.606607] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.611736] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.611736] env[63175]: value = "task-1248272" [ 884.611736] env[63175]: _type = "Task" [ 884.611736] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.619599] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248272, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.649859] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248271, 'name': ReconfigVM_Task, 'duration_secs': 0.591115} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.650251] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfigured VM instance instance-0000004d to attach disk [datastore1] a4304579-f829-433e-a878-1050fac08c2e/a4304579-f829-433e-a878-1050fac08c2e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.650999] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f26646a-9c19-41b6-9cc8-8ad540b364ea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.658711] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 884.658711] env[63175]: value = "task-1248273" [ 884.658711] env[63175]: _type = "Task" [ 884.658711] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.668138] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248273, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.701628] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.701937] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.702201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.702362] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.702543] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.705010] env[63175]: INFO nova.compute.manager [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Terminating instance [ 884.771821] env[63175]: DEBUG oslo_vmware.api [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52485973-5351-5f03-1206-8fcf40a6fd24, 'name': SearchDatastore_Task, 'duration_secs': 0.013813} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.772141] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.795475] env[63175]: INFO nova.compute.manager [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Took 21.72 seconds to build instance. [ 884.934019] env[63175]: DEBUG nova.compute.manager [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 884.934283] env[63175]: DEBUG nova.compute.manager [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing instance network info cache due to event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 884.934539] env[63175]: DEBUG oslo_concurrency.lockutils [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.934692] env[63175]: DEBUG oslo_concurrency.lockutils [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.934871] env[63175]: DEBUG nova.network.neutron [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.066929] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "refresh_cache-f37d34d4-b13e-4ce0-b010-3a06feab6324" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.067151] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquired lock "refresh_cache-f37d34d4-b13e-4ce0-b010-3a06feab6324" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.067332] env[63175]: DEBUG nova.network.neutron [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.079548] env[63175]: DEBUG nova.scheduler.client.report [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 885.122197] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248272, 'name': CreateVM_Task, 'duration_secs': 0.376468} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.122410] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.123667] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.123667] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.123856] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.124227] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f2062f7-1aab-4432-a518-50d76612c84e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.134586] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 885.134586] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5256f6d8-ec4f-08ae-2456-20f6f1183bee" [ 885.134586] env[63175]: _type = "Task" [ 885.134586] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.146952] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5256f6d8-ec4f-08ae-2456-20f6f1183bee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.170216] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248273, 'name': Rename_Task, 'duration_secs': 0.199055} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.170495] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.170751] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27728500-df65-4a88-86a2-e8567d84d943 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.179219] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 885.179219] env[63175]: value = "task-1248274" [ 885.179219] env[63175]: _type = "Task" [ 885.179219] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.189326] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.193622] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.194554] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d0063e-0fbe-4c68-93e6-b5cdea577d32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.202636] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.202921] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c9ea8db-54ec-4ed2-9345-52a072f7d37b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.209566] env[63175]: DEBUG nova.compute.manager [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 885.209788] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.210615] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c677cd-9995-40f0-8d06-fc9264bf436a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.218972] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.219252] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e294afa1-8850-4144-88db-712fef4224d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.227399] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 885.227399] env[63175]: value = "task-1248276" [ 885.227399] env[63175]: _type = "Task" [ 885.227399] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.236747] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248276, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.297692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8ee514e0-8080-48c6-aff6-14231fe5b97e tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.232s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.302052] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.302354] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.302534] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleting the datastore file [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.302812] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f934fa2-86f8-4e62-a699-7fda14e86cd0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.309935] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 885.309935] env[63175]: value = "task-1248277" [ 885.309935] env[63175]: _type = "Task" [ 885.309935] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.318395] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.583839] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.584396] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 885.590405] env[63175]: DEBUG nova.network.neutron [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.592267] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.820s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.649558] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5256f6d8-ec4f-08ae-2456-20f6f1183bee, 'name': SearchDatastore_Task, 'duration_secs': 0.016313} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.650263] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.650515] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.650982] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.651206] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.651482] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.651704] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66b15943-72c1-48e3-a82c-e7226b474b9e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.661667] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.661829] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.662607] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-814df069-b663-4571-9179-c20d9db9c4a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.671591] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 885.671591] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5200559e-35c5-b630-8eb1-08dba5d2a7d9" [ 885.671591] env[63175]: _type = "Task" [ 885.671591] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.681119] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5200559e-35c5-b630-8eb1-08dba5d2a7d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.690957] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248274, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.738511] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248276, 'name': PowerOffVM_Task, 'duration_secs': 0.385946} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.738925] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.739147] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.739474] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b998a5c-cdb7-4995-8b07-ce526fa778d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.782101] env[63175]: DEBUG nova.network.neutron [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updated VIF entry in instance network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.782599] env[63175]: DEBUG nova.network.neutron [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.820709] env[63175]: DEBUG oslo_vmware.api [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241215} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.821130] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.821225] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.821433] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.846990] env[63175]: INFO nova.scheduler.client.report [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted allocations for instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f [ 885.899445] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.899711] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.899873] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleting the datastore file [datastore2] 5d25451b-69aa-4af7-be71-451fa0ac1e39 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.900188] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d51b04c-2fb3-4e8e-a5b7-2411718b370d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.909371] env[63175]: DEBUG nova.network.neutron [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.912037] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for the task: (returnval){ [ 885.912037] env[63175]: value = "task-1248279" [ 885.912037] env[63175]: _type = "Task" [ 885.912037] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.922986] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.096157] env[63175]: DEBUG nova.compute.utils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.100646] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 886.101123] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 886.166030] env[63175]: DEBUG nova.policy [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.182948] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5200559e-35c5-b630-8eb1-08dba5d2a7d9, 'name': SearchDatastore_Task, 'duration_secs': 0.013954} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.189299] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53dfcd42-b1da-4ea8-b921-0fa55009a8ed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.195505] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 886.195505] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5228d1d0-5d62-504a-b74c-576f70cf7123" [ 886.195505] env[63175]: _type = "Task" [ 886.195505] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.198666] env[63175]: DEBUG oslo_vmware.api [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248274, 'name': PowerOnVM_Task, 'duration_secs': 0.881371} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.201756] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.201980] env[63175]: INFO nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Took 8.62 seconds to spawn the instance on the hypervisor. [ 886.202178] env[63175]: DEBUG nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.205287] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a467d94-cf7d-41cc-a1b5-12ffaa0b8f23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.214451] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5228d1d0-5d62-504a-b74c-576f70cf7123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.285416] env[63175]: DEBUG oslo_concurrency.lockutils [req-3b415180-0290-4055-8a10-6b11160668b3 req-8608edcb-03c2-498e-9093-0b5144b74853 service nova] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.307195] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70979e1-652b-4ca1-92cc-e7c823e4228d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.315822] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97494db-4853-4b84-b3d7-9e76e0fec890 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.345993] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1038cd5-58b7-40c9-ba46-edf8ba3f535b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.353773] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.355161] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4787f51-747b-4a89-92ec-2f4060a38433 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.370920] env[63175]: DEBUG nova.compute.provider_tree [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.413811] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Releasing lock "refresh_cache-f37d34d4-b13e-4ce0-b010-3a06feab6324" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.414379] env[63175]: DEBUG nova.compute.manager [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 886.414623] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.415657] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f0bde0-e575-4c62-84f8-dba2a27af3b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.427992] env[63175]: DEBUG oslo_vmware.api [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Task: {'id': task-1248279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.38219} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.430243] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.430442] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.430627] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.430805] env[63175]: INFO nova.compute.manager [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Took 1.22 seconds to destroy the instance on the hypervisor. [ 886.431095] env[63175]: DEBUG oslo.service.loopingcall [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.431338] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.432180] env[63175]: DEBUG nova.compute.manager [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 886.432284] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.433907] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4840a555-472b-4a0e-afd1-bed0c95cc569 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.442749] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 886.442749] env[63175]: value = "task-1248280" [ 886.442749] env[63175]: _type = "Task" [ 886.442749] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.453283] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.576817] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 886.577672] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5691188d-473b-4352-a7cb-af5115d0f109 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.586825] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 886.588039] env[63175]: ERROR oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk due to incomplete transfer. [ 886.588039] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b7bca8d9-fb71-42e7-8759-80db805f0a10 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.595928] env[63175]: DEBUG oslo_vmware.rw_handles [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526c25e7-12e2-558c-a17f-11075c88ca8e/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 886.596181] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Uploaded image 64e4e8b2-7861-480b-b338-14289e25bd77 to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 886.598353] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 886.598639] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-86e64790-66b2-4b00-8ac3-580a29e693bd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.600763] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 886.610278] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 886.610278] env[63175]: value = "task-1248281" [ 886.610278] env[63175]: _type = "Task" [ 886.610278] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.621504] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248281, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.729070] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5228d1d0-5d62-504a-b74c-576f70cf7123, 'name': SearchDatastore_Task, 'duration_secs': 0.022442} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.729919] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.730069] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0645fffe-9910-43af-af40-126592aefaff/0645fffe-9910-43af-af40-126592aefaff.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.730425] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8272b2a-ce08-4268-9509-038a370e97be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.739886] env[63175]: INFO nova.compute.manager [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Took 23.37 seconds to build instance. [ 886.745902] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 886.745902] env[63175]: value = "task-1248282" [ 886.745902] env[63175]: _type = "Task" [ 886.745902] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.752740] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Successfully created port: 8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.758926] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248282, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.877024] env[63175]: DEBUG nova.scheduler.client.report [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 886.938942] env[63175]: DEBUG nova.compute.manager [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Received event network-changed-fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 886.939633] env[63175]: DEBUG nova.compute.manager [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Refreshing instance network info cache due to event network-changed-fcbca53b-aabe-41c3-bfdb-0821599b70d8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 886.940153] env[63175]: DEBUG oslo_concurrency.lockutils [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] Acquiring lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.940465] env[63175]: DEBUG oslo_concurrency.lockutils [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] Acquired lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.940801] env[63175]: DEBUG nova.network.neutron [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Refreshing network info cache for port fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.954694] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248280, 'name': PowerOffVM_Task, 'duration_secs': 0.176621} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.956351] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.956558] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.957616] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e03400a1-a847-418d-b8bc-868d270609cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.990415] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.990642] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.991619] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleting the datastore file [datastore2] f37d34d4-b13e-4ce0-b010-3a06feab6324 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.991619] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55d27913-82e1-4000-a59c-50ea94e0a89b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.000207] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for the task: (returnval){ [ 887.000207] env[63175]: value = "task-1248284" [ 887.000207] env[63175]: _type = "Task" [ 887.000207] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.010159] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248284, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.134028] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248281, 'name': Destroy_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.175717] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-vif-unplugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.175974] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.176202] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.176385] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.176550] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] No waiting events found dispatching network-vif-unplugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 887.176723] env[63175]: WARNING nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received unexpected event network-vif-unplugged-6a69f7df-439f-4044-8394-305f9f3a93bb for instance with vm_state shelved_offloaded and task_state None. [ 887.176969] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.177699] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing instance network info cache due to event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 887.177970] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.178149] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.178328] env[63175]: DEBUG nova.network.neutron [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.242536] env[63175]: DEBUG oslo_concurrency.lockutils [None req-7c3d9847-0f4d-42ec-8823-ab4859349832 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.881s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.259526] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248282, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.512096] env[63175]: DEBUG oslo_vmware.api [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Task: {'id': task-1248284, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26154} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.512608] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.512781] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.512967] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.513217] env[63175]: INFO nova.compute.manager [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Took 1.10 seconds to destroy the instance on the hypervisor. [ 887.513487] env[63175]: DEBUG oslo.service.loopingcall [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.514361] env[63175]: DEBUG nova.compute.manager [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 887.517482] env[63175]: DEBUG nova.network.neutron [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.535519] env[63175]: DEBUG nova.network.neutron [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.570126] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.622386] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 887.624347] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248281, 'name': Destroy_Task, 'duration_secs': 0.523651} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.624818] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Destroyed the VM [ 887.625070] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 887.625408] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a81f050b-ee1e-4eec-83f8-4f2cc2add401 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.632711] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 887.632711] env[63175]: value = "task-1248285" [ 887.632711] env[63175]: _type = "Task" [ 887.632711] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.640671] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248285, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.648599] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.648842] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.649019] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.649223] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.649503] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.649758] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.649992] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.650212] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.650404] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.650579] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.650789] env[63175]: DEBUG nova.virt.hardware [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.651881] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3acf84-bc7a-4699-a2f1-5082f7b17863 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.659802] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8443e54-a12c-41dd-a3df-f14a49e94287 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.756686] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248282, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719374} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.759190] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0645fffe-9910-43af-af40-126592aefaff/0645fffe-9910-43af-af40-126592aefaff.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.759448] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.759729] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ded94928-5fba-47ec-9838-5d44411617d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.768194] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 887.768194] env[63175]: value = "task-1248286" [ 887.768194] env[63175]: _type = "Task" [ 887.768194] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.777749] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.858570] env[63175]: DEBUG nova.network.neutron [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updated VIF entry in instance network info cache for port fcbca53b-aabe-41c3-bfdb-0821599b70d8. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.858964] env[63175]: DEBUG nova.network.neutron [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updating instance_info_cache with network_info: [{"id": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "address": "fa:16:3e:4a:fe:51", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbca53b-aa", "ovs_interfaceid": "fcbca53b-aabe-41c3-bfdb-0821599b70d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.888333] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.296s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.894082] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.540s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.894705] env[63175]: DEBUG nova.objects.instance [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lazy-loading 'resources' on Instance uuid 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.953317] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.953583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.953803] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.953993] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.954201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.956267] env[63175]: INFO nova.compute.manager [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Terminating instance [ 888.039009] env[63175]: DEBUG nova.network.neutron [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.043279] env[63175]: DEBUG nova.network.neutron [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updated VIF entry in instance network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.043279] env[63175]: DEBUG nova.network.neutron [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6a69f7df-43", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.073184] env[63175]: INFO nova.compute.manager [-] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Took 1.64 seconds to deallocate network for instance. [ 888.145145] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248285, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.278440] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127337} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.278737] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.279541] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8de387d-e846-423c-aeb5-4655d536c287 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.303719] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 0645fffe-9910-43af-af40-126592aefaff/0645fffe-9910-43af-af40-126592aefaff.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.303994] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ca20548-d1bb-42fd-ba75-5300a7514d98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.327488] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 888.327488] env[63175]: value = "task-1248287" [ 888.327488] env[63175]: _type = "Task" [ 888.327488] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.335931] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248287, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.361806] env[63175]: DEBUG oslo_concurrency.lockutils [req-fa39bac7-4787-4990-9792-3ecfa0033f5b req-1f61d001-1c34-414a-af39-76836635c40f service nova] Releasing lock "refresh_cache-9b72d607-be0f-4caa-b9e2-b25bbbe30f54" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.400617] env[63175]: DEBUG nova.objects.instance [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lazy-loading 'numa_topology' on Instance uuid 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.460318] env[63175]: DEBUG nova.compute.manager [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 888.460911] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.461599] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8dc40c-4dd7-484a-93ba-b9b92c1ba390 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.476060] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.476060] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12eef791-de4e-459f-88fe-77bae6052975 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.482118] env[63175]: INFO nova.scheduler.client.report [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocation for migration 2a22dd23-126d-4a3c-b064-3e2402ed6681 [ 888.541917] env[63175]: INFO nova.compute.manager [-] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Took 1.03 seconds to deallocate network for instance. [ 888.544790] env[63175]: DEBUG oslo_concurrency.lockutils [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.545371] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-vif-deleted-dc3362cd-7410-4a9d-ba7e-35caad1458fb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 888.545695] env[63175]: INFO nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Neutron deleted interface dc3362cd-7410-4a9d-ba7e-35caad1458fb; detaching it from the instance and deleting it from the info cache [ 888.546048] env[63175]: DEBUG nova.network.neutron [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [{"id": "cba26230-a8fc-4ce1-8729-feb6678721e7", "address": "fa:16:3e:92:70:8b", "network": {"id": "a11a1a05-e5f8-460b-9c02-7d20847bd84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1360242438", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.253", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "975415f67c2645678f0815424128063b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcba26230-a8", "ovs_interfaceid": "cba26230-a8fc-4ce1-8729-feb6678721e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.583509] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.645975] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248285, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.665633] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Successfully updated port: 8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.724862] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.725409] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.841595] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248287, 'name': ReconfigVM_Task, 'duration_secs': 0.326042} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.842228] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 0645fffe-9910-43af-af40-126592aefaff/0645fffe-9910-43af-af40-126592aefaff.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.843270] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fca6121-8e7c-4299-b097-07078ccae4ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.851334] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 888.851334] env[63175]: value = "task-1248289" [ 888.851334] env[63175]: _type = "Task" [ 888.851334] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.861443] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248289, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.904898] env[63175]: DEBUG nova.objects.base [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Object Instance<5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f> lazy-loaded attributes: resources,numa_topology {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 888.979398] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-changed-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 888.980436] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing instance network info cache due to event network-changed-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 888.980436] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.980436] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.980436] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing network info cache for port 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.989458] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0818c226-9345-4758-a9a1-13c378623223 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.698s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.047815] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.048174] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.048392] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore1] 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.048911] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab15caf7-ea07-45ca-9ea0-a69b491f88ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.053931] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c2cf281-42cf-4b74-bb45-e7f399d51aaf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.057091] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.059517] env[63175]: DEBUG oslo_vmware.api [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 889.059517] env[63175]: value = "task-1248290" [ 889.059517] env[63175]: _type = "Task" [ 889.059517] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.069597] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825040e2-6151-4760-9b6c-f1df23023846 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.085809] env[63175]: DEBUG oslo_vmware.api [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.096632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1a9178-33cb-48ff-8e09-a35a4c4660e6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.107875] env[63175]: DEBUG nova.compute.manager [req-add20be2-123a-443f-ade1-4d5355a0ae3b req-f9d11c96-8e43-47c0-aaca-089f4f660f13 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Detach interface failed, port_id=dc3362cd-7410-4a9d-ba7e-35caad1458fb, reason: Instance 5d25451b-69aa-4af7-be71-451fa0ac1e39 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 889.113615] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a90ba5-e497-45ba-a502-3987ed5281f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.149399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257e25cd-feae-45bc-a75b-bde88ab84a53 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.161341] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dda1ad-4ca1-4f6e-9c06-6c0f19e42bb0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.165353] env[63175]: DEBUG oslo_vmware.api [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248285, 'name': RemoveSnapshot_Task, 'duration_secs': 1.189422} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.165630] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 889.165893] env[63175]: INFO nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Took 16.23 seconds to snapshot the instance on the hypervisor. [ 889.169121] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.169167] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.169366] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.179241] env[63175]: DEBUG nova.compute.provider_tree [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.201168] env[63175]: DEBUG nova.compute.manager [req-d2a040a4-63b7-415d-9315-eac29f24e711 req-f665a237-3896-4085-b480-26efd1b0c668 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Received event network-vif-deleted-cba26230-a8fc-4ce1-8729-feb6678721e7 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 889.201168] env[63175]: INFO nova.compute.manager [req-d2a040a4-63b7-415d-9315-eac29f24e711 req-f665a237-3896-4085-b480-26efd1b0c668 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Neutron deleted interface cba26230-a8fc-4ce1-8729-feb6678721e7; detaching it from the instance and deleting it from the info cache [ 889.201168] env[63175]: DEBUG nova.network.neutron [req-d2a040a4-63b7-415d-9315-eac29f24e711 req-f665a237-3896-4085-b480-26efd1b0c668 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.229900] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 889.362060] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248289, 'name': Rename_Task, 'duration_secs': 0.144679} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.362279] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.362538] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-628bd20e-a3b5-4b59-9378-3f658442487e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.369490] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 889.369490] env[63175]: value = "task-1248291" [ 889.369490] env[63175]: _type = "Task" [ 889.369490] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.378269] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248291, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.569925] env[63175]: DEBUG oslo_vmware.api [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.497534} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.570470] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.570912] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.571076] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.571204] env[63175]: INFO nova.compute.manager [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Took 1.11 seconds to destroy the instance on the hypervisor. [ 889.571492] env[63175]: DEBUG oslo.service.loopingcall [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.571725] env[63175]: DEBUG nova.compute.manager [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 889.571853] env[63175]: DEBUG nova.network.neutron [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.671644] env[63175]: DEBUG nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance disappeared during snapshot {{(pid=63175) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 889.683499] env[63175]: DEBUG nova.scheduler.client.report [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 889.687412] env[63175]: DEBUG nova.compute.manager [None req-80282c8e-a2c3-49c9-a757-31f2d7564908 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image not found during clean up 64e4e8b2-7861-480b-b338-14289e25bd77 {{(pid=63175) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 889.707558] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37d9079c-99ad-4d0d-a62a-1a3f74aa6d44 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.720855] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4717ad-d215-42ac-acfd-a4cc358ac795 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.762327] env[63175]: DEBUG nova.compute.manager [req-d2a040a4-63b7-415d-9315-eac29f24e711 req-f665a237-3896-4085-b480-26efd1b0c668 service nova] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Detach interface failed, port_id=cba26230-a8fc-4ce1-8729-feb6678721e7, reason: Instance 5d25451b-69aa-4af7-be71-451fa0ac1e39 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 889.763849] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.764633] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.835460] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.841362] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updated VIF entry in instance network info cache for port 3cd631a8-97d6-4ff8-87c6-ab78b6d39c51. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.841769] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.882705] env[63175]: DEBUG oslo_vmware.api [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248291, 'name': PowerOnVM_Task, 'duration_secs': 0.445145} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.884090] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.884420] env[63175]: INFO nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Took 8.12 seconds to spawn the instance on the hypervisor. [ 889.884726] env[63175]: DEBUG nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 889.888440] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de067b16-4ac9-492f-acd7-85a4fdf84926 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.024370] env[63175]: DEBUG nova.network.neutron [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Updating instance_info_cache with network_info: [{"id": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "address": "fa:16:3e:7d:95:a4", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ffc5202-b3", "ovs_interfaceid": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.192689] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.299s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.195198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.612s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.195436] env[63175]: DEBUG nova.objects.instance [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lazy-loading 'resources' on Instance uuid 5d25451b-69aa-4af7-be71-451fa0ac1e39 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.344757] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.345050] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Received event network-vif-plugged-8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 890.345262] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Acquiring lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.345490] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.345655] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.345831] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] No waiting events found dispatching network-vif-plugged-8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 890.346015] env[63175]: WARNING nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Received unexpected event network-vif-plugged-8ffc5202-b3e9-45c5-9821-2b18e73a48e8 for instance with vm_state building and task_state spawning. [ 890.346198] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Received event network-changed-8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 890.346364] env[63175]: DEBUG nova.compute.manager [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Refreshing instance network info cache due to event network-changed-8ffc5202-b3e9-45c5-9821-2b18e73a48e8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 890.346537] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Acquiring lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.392656] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.392882] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.393132] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.393334] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.393510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.395646] env[63175]: INFO nova.compute.manager [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Terminating instance [ 890.409217] env[63175]: INFO nova.compute.manager [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Took 22.06 seconds to build instance. [ 890.439468] env[63175]: DEBUG nova.network.neutron [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.526773] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.527119] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Instance network_info: |[{"id": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "address": "fa:16:3e:7d:95:a4", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ffc5202-b3", "ovs_interfaceid": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 890.527803] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Acquired lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.527803] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Refreshing network info cache for port 8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.530675] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:95:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ffc5202-b3e9-45c5-9821-2b18e73a48e8', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.538220] env[63175]: DEBUG oslo.service.loopingcall [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.538526] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.538752] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-812bb459-c907-4639-8e9b-373637acff69 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.560220] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.560220] env[63175]: value = "task-1248292" [ 890.560220] env[63175]: _type = "Task" [ 890.560220] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.576645] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248292, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.703171] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5929f29e-1c86-4bc3-a99d-e756f096cfc8 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.547s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.704101] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.869s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.704350] env[63175]: INFO nova.compute.manager [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Unshelving [ 890.825693] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.825953] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.900039] env[63175]: DEBUG nova.compute.manager [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 890.900039] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.900338] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce9989a-f327-4654-9cae-e668f4772159 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.905959] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a288a98-35ee-400c-b97c-a1ee21486e4f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.911059] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d53a35c8-a502-40bb-b6e9-4e1d8be5237d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.573s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.911345] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.911967] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-707eb8f3-96ab-490f-9e13-b7efa7131eca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.916866] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b75b70-eaeb-47cb-8e0b-ed3686264260 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.921147] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 890.921147] env[63175]: value = "task-1248293" [ 890.921147] env[63175]: _type = "Task" [ 890.921147] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.949770] env[63175]: INFO nova.compute.manager [-] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Took 1.38 seconds to deallocate network for instance. [ 890.952915] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da3a112-7ff1-4b70-a8fc-623a704654e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.961084] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248293, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.966880] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3ed99f-7d3e-4d75-ba8b-c347cc6fc317 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.987725] env[63175]: DEBUG nova.compute.provider_tree [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.071931] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248292, 'name': CreateVM_Task, 'duration_secs': 0.360924} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.072443] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.072979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.073348] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.073634] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.073935] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55d973b2-1b5a-423a-bcad-fdc22c6acae2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.083800] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 891.083800] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b932a-3252-dda8-3784-84671ac575c2" [ 891.083800] env[63175]: _type = "Task" [ 891.083800] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.091868] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b932a-3252-dda8-3784-84671ac575c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.329796] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 891.393338] env[63175]: DEBUG nova.compute.manager [req-1e8c8ae3-60a4-47df-9df2-b85c0182526c req-cd500347-2660-4942-9d9e-70b752a70ec9 service nova] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Received event network-vif-deleted-e68b5b69-c8d6-47f5-97a8-96db33cb4eb4 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 891.432060] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248293, 'name': PowerOffVM_Task, 'duration_secs': 0.258821} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.432335] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.433020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.433020] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9d3d274-05f4-43ba-8725-90d8b7f93978 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.463800] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.494241] env[63175]: DEBUG nova.scheduler.client.report [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 891.550541] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Updated VIF entry in instance network info cache for port 8ffc5202-b3e9-45c5-9821-2b18e73a48e8. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.550918] env[63175]: DEBUG nova.network.neutron [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Updating instance_info_cache with network_info: [{"id": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "address": "fa:16:3e:7d:95:a4", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ffc5202-b3", "ovs_interfaceid": "8ffc5202-b3e9-45c5-9821-2b18e73a48e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.553473] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.553676] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.553862] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleting the datastore file [datastore2] 4bd91412-c41b-41a1-a648-6b905d826ee3 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.554375] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc299a42-9284-47dc-bd18-66ebeb50887b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.563541] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 891.563541] env[63175]: value = "task-1248295" [ 891.563541] env[63175]: _type = "Task" [ 891.563541] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.573244] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248295, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.592960] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521b932a-3252-dda8-3784-84671ac575c2, 'name': SearchDatastore_Task, 'duration_secs': 0.024274} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.593344] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.593620] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.593911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.594169] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.594404] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.594703] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a5a5e29-d323-4d8c-8d9b-96a87dcf73ef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.603441] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.603671] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.604434] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3578d8fa-0685-4e34-a6cd-5a8a80d09cb0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.610184] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 891.610184] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52883476-ef31-513b-43ad-c0dcb46fb11c" [ 891.610184] env[63175]: _type = "Task" [ 891.610184] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.618338] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52883476-ef31-513b-43ad-c0dcb46fb11c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.725532] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.853062] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.945472] env[63175]: DEBUG nova.compute.manager [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 891.945646] env[63175]: DEBUG nova.compute.manager [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing instance network info cache due to event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 891.945989] env[63175]: DEBUG oslo_concurrency.lockutils [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.945989] env[63175]: DEBUG oslo_concurrency.lockutils [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.946209] env[63175]: DEBUG nova.network.neutron [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.002031] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.804s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.002031] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.945s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.002660] env[63175]: DEBUG nova.objects.instance [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lazy-loading 'resources' on Instance uuid f37d34d4-b13e-4ce0-b010-3a06feab6324 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.029269] env[63175]: INFO nova.scheduler.client.report [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Deleted allocations for instance 5d25451b-69aa-4af7-be71-451fa0ac1e39 [ 892.055234] env[63175]: DEBUG oslo_concurrency.lockutils [req-e5b6f55a-c5e9-4f3f-ade1-c1848cea93d9 req-460a8133-9664-4c8d-baad-4bbfcae656f2 service nova] Releasing lock "refresh_cache-8d3bd040-fa87-44b8-a6c1-fe6be21b0415" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.080803] env[63175]: DEBUG oslo_vmware.api [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248295, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198634} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.080803] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.080803] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.080803] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.080803] env[63175]: INFO nova.compute.manager [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Took 1.18 seconds to destroy the instance on the hypervisor. [ 892.081267] env[63175]: DEBUG oslo.service.loopingcall [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.082425] env[63175]: DEBUG nova.compute.manager [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 892.082530] env[63175]: DEBUG nova.network.neutron [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.121649] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52883476-ef31-513b-43ad-c0dcb46fb11c, 'name': SearchDatastore_Task, 'duration_secs': 0.009749} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.122483] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c8cc2f-50bb-4043-bdba-f27e11d9f32e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.129327] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 892.129327] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f75563-75f3-6884-0bd0-54c689ee2f86" [ 892.129327] env[63175]: _type = "Task" [ 892.129327] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.138293] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f75563-75f3-6884-0bd0-54c689ee2f86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.543505] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5a1f6b9e-da09-4eed-ac21-08bcd2c4b593 tempest-ServersTestMultiNic-1494322380 tempest-ServersTestMultiNic-1494322380-project-member] Lock "5d25451b-69aa-4af7-be71-451fa0ac1e39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.841s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.643497] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f75563-75f3-6884-0bd0-54c689ee2f86, 'name': SearchDatastore_Task, 'duration_secs': 0.045312} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.643990] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.644402] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 8d3bd040-fa87-44b8-a6c1-fe6be21b0415/8d3bd040-fa87-44b8-a6c1-fe6be21b0415.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.644805] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-593f9db4-47e4-4f11-80a3-20a49cc59308 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.654655] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 892.654655] env[63175]: value = "task-1248296" [ 892.654655] env[63175]: _type = "Task" [ 892.654655] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.664990] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.691731] env[63175]: DEBUG nova.network.neutron [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updated VIF entry in instance network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.691998] env[63175]: DEBUG nova.network.neutron [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.713632] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c42fe54-26cc-4504-af78-0ae6e58fcbdc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.723008] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b49e325-18a6-4538-b2c1-4b5e94fe2b1b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.754827] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd88d39-f01a-4057-b201-26eafad1f5d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.762834] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe367c7c-ff76-43ac-8c5b-573c6bfd08e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.777545] env[63175]: DEBUG nova.compute.provider_tree [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.872244] env[63175]: DEBUG nova.network.neutron [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.166041] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248296, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.195190] env[63175]: DEBUG oslo_concurrency.lockutils [req-cdce17db-5306-435f-90c6-0cc7110cfeeb req-ca194ee8-e4af-4345-a179-9923fdee3f60 service nova] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.281680] env[63175]: DEBUG nova.scheduler.client.report [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 893.375653] env[63175]: INFO nova.compute.manager [-] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Took 1.29 seconds to deallocate network for instance. [ 893.422137] env[63175]: DEBUG nova.compute.manager [req-d92b9cf2-3cde-4a00-bc79-6c47f2a26ea1 req-0716c986-85c1-4324-90d0-3038fdf9cc1c service nova] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Received event network-vif-deleted-a30697ac-b3b6-46c0-acb8-9b3ff5fa3aaf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 893.666032] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.862797} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.667032] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 8d3bd040-fa87-44b8-a6c1-fe6be21b0415/8d3bd040-fa87-44b8-a6c1-fe6be21b0415.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.667032] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.667198] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce2cbf71-82c9-4a85-a306-bcf301fa6a4b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.674957] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 893.674957] env[63175]: value = "task-1248297" [ 893.674957] env[63175]: _type = "Task" [ 893.674957] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.683902] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.788268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.790827] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.027s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.792485] env[63175]: INFO nova.compute.claims [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.817031] env[63175]: INFO nova.scheduler.client.report [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Deleted allocations for instance f37d34d4-b13e-4ce0-b010-3a06feab6324 [ 893.883773] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.185252] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061664} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.185601] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.186352] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab4f907-1baa-4d76-99c7-45c507503677 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.210654] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 8d3bd040-fa87-44b8-a6c1-fe6be21b0415/8d3bd040-fa87-44b8-a6c1-fe6be21b0415.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.210999] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef174db7-0d91-4f74-85c0-717f99cd0e9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.230900] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 894.230900] env[63175]: value = "task-1248298" [ 894.230900] env[63175]: _type = "Task" [ 894.230900] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.241414] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248298, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.327244] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0f9a9105-4547-4d10-8e58-28f171484f1b tempest-ServerShowV247Test-1112214181 tempest-ServerShowV247Test-1112214181-project-member] Lock "f37d34d4-b13e-4ce0-b010-3a06feab6324" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.774s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.741592] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248298, 'name': ReconfigVM_Task, 'duration_secs': 0.32222} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.741852] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 8d3bd040-fa87-44b8-a6c1-fe6be21b0415/8d3bd040-fa87-44b8-a6c1-fe6be21b0415.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.744506] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51274a6a-b6d0-4fae-b1c5-3e49d007e48b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.752040] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 894.752040] env[63175]: value = "task-1248299" [ 894.752040] env[63175]: _type = "Task" [ 894.752040] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.767451] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248299, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.001214] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59df77c4-afcd-4edd-a71d-c5c71a0922aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.010576] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b763277-74e0-4f39-b80e-e39374db5525 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.056018] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1860a6d2-0863-42a6-87b2-4b62bd085c23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.066088] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f412adc-4ba7-4d74-952c-a56272dee5d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.080521] env[63175]: DEBUG nova.compute.provider_tree [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.264589] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248299, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.583871] env[63175]: DEBUG nova.scheduler.client.report [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 895.762884] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248299, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.088196] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.088746] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 896.091341] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.628s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.091560] env[63175]: DEBUG nova.objects.instance [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'resources' on Instance uuid 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.263855] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248299, 'name': Rename_Task, 'duration_secs': 1.152602} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.264157] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.264432] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3de0b0f2-6b39-486c-973f-1326c8c4d5bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.271615] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 896.271615] env[63175]: value = "task-1248300" [ 896.271615] env[63175]: _type = "Task" [ 896.271615] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.279827] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.597250] env[63175]: DEBUG nova.compute.utils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.599025] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 896.599025] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.642106] env[63175]: DEBUG nova.policy [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e43774590162406a981cf0d5e20f5a1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83fa8f008a48484195c88ed76ceece0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.782542] env[63175]: DEBUG oslo_vmware.api [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248300, 'name': PowerOnVM_Task, 'duration_secs': 0.510917} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.785190] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.785429] env[63175]: INFO nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Took 9.16 seconds to spawn the instance on the hypervisor. [ 896.785639] env[63175]: DEBUG nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 896.786628] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fc7444-7b2e-4f84-8ce8-3c07700fb84a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.832726] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed52e6dc-82eb-430e-9b92-5c5a371817e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.842111] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cee4de-e5e2-463a-b73c-32156055d338 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.874216] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141f88a8-71d3-49b8-a957-bc7e83127dd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.882553] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865eed48-d761-4fb0-bab4-870820ddd7dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.897271] env[63175]: DEBUG nova.compute.provider_tree [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.010543] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Successfully created port: eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.102308] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 897.307352] env[63175]: INFO nova.compute.manager [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Took 14.02 seconds to build instance. [ 897.400509] env[63175]: DEBUG nova.scheduler.client.report [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 897.809296] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ac4a3325-fdad-467c-991c-c142b445192a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.530s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.906358] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.815s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.910774] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.185s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.910896] env[63175]: DEBUG nova.objects.instance [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lazy-loading 'pci_requests' on Instance uuid 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.932693] env[63175]: INFO nova.scheduler.client.report [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted allocations for instance 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68 [ 898.043335] env[63175]: DEBUG oslo_concurrency.lockutils [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.043621] env[63175]: DEBUG oslo_concurrency.lockutils [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.114243] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 898.139372] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.139691] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.139861] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.140066] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.140227] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.140383] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.140596] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.140800] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.140987] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.141185] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.141369] env[63175]: DEBUG nova.virt.hardware [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.142645] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101de1e9-4340-49b1-a8cc-6a86f33f6ed3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.151678] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf2534-2865-4e55-b5b4-8a4f63333f9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.379350] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.380217] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.380217] env[63175]: DEBUG nova.compute.manager [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 898.380775] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c627e77b-0d4d-40bf-aa04-e735a3477577 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.388829] env[63175]: DEBUG nova.compute.manager [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63175) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 898.389414] env[63175]: DEBUG nova.objects.instance [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'flavor' on Instance uuid 8d3bd040-fa87-44b8-a6c1-fe6be21b0415 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.392412] env[63175]: DEBUG nova.compute.manager [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Received event network-vif-plugged-eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 898.392566] env[63175]: DEBUG oslo_concurrency.lockutils [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] Acquiring lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.392766] env[63175]: DEBUG oslo_concurrency.lockutils [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.392933] env[63175]: DEBUG oslo_concurrency.lockutils [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.393119] env[63175]: DEBUG nova.compute.manager [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] No waiting events found dispatching network-vif-plugged-eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 898.393337] env[63175]: WARNING nova.compute.manager [req-11c56b89-c803-4bd1-8134-cbf59ac4cd4c req-ae089590-e800-4309-b7a9-ff1235b90974 service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Received unexpected event network-vif-plugged-eaec69ba-7e1a-4abf-b799-33836e2e6c04 for instance with vm_state building and task_state spawning. [ 898.414917] env[63175]: DEBUG nova.objects.instance [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lazy-loading 'numa_topology' on Instance uuid 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.439789] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffb1745e-2b8b-4917-907b-e8f808fa8de5 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "1b8b3fdf-642b-40cc-a8ba-07ecd5519a68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.486s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.485767] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Successfully updated port: eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.547604] env[63175]: INFO nova.compute.manager [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Detaching volume 37c89c77-c578-4a98-b68a-e8d493df963d [ 898.581740] env[63175]: INFO nova.virt.block_device [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Attempting to driver detach volume 37c89c77-c578-4a98-b68a-e8d493df963d from mountpoint /dev/sdb [ 898.582076] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 898.582324] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269035', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'name': 'volume-37c89c77-c578-4a98-b68a-e8d493df963d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3', 'attached_at': '', 'detached_at': '', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'serial': '37c89c77-c578-4a98-b68a-e8d493df963d'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 898.583302] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e9ed19-dbd6-46a3-be32-c9ca5772b871 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.605311] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9774fd-8f67-4904-aa17-2361375f6b30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.612515] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa47c2b-a74c-4a2f-8074-af2692f6bb3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.632851] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440e665d-6db2-4a24-bc2b-344133fed9c9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.649078] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] The volume has not been displaced from its original location: [datastore2] volume-37c89c77-c578-4a98-b68a-e8d493df963d/volume-37c89c77-c578-4a98-b68a-e8d493df963d.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 898.654327] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 898.654910] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba869b62-6ca3-445c-8674-a1f75e56f6a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.672912] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 898.672912] env[63175]: value = "task-1248301" [ 898.672912] env[63175]: _type = "Task" [ 898.672912] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.680690] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248301, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.917401] env[63175]: INFO nova.compute.claims [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.991732] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.991732] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.991732] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.186049] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248301, 'name': ReconfigVM_Task, 'duration_secs': 0.286511} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.186049] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 899.188152] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1038f6ba-a08f-4792-bd6d-326db9c21662 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.204655] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 899.204655] env[63175]: value = "task-1248302" [ 899.204655] env[63175]: _type = "Task" [ 899.204655] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.213326] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248302, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.399254] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.399631] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9e29257-aa25-4491-b841-6d4e7d82bd19 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.407436] env[63175]: DEBUG oslo_vmware.api [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 899.407436] env[63175]: value = "task-1248303" [ 899.407436] env[63175]: _type = "Task" [ 899.407436] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.417830] env[63175]: DEBUG oslo_vmware.api [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.523741] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.656483] env[63175]: DEBUG nova.network.neutron [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Updating instance_info_cache with network_info: [{"id": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "address": "fa:16:3e:b1:53:28", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaec69ba-7e", "ovs_interfaceid": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.715198] env[63175]: DEBUG oslo_vmware.api [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248302, 'name': ReconfigVM_Task, 'duration_secs': 0.141046} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.715558] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269035', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'name': 'volume-37c89c77-c578-4a98-b68a-e8d493df963d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3', 'attached_at': '', 'detached_at': '', 'volume_id': '37c89c77-c578-4a98-b68a-e8d493df963d', 'serial': '37c89c77-c578-4a98-b68a-e8d493df963d'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 899.917730] env[63175]: DEBUG oslo_vmware.api [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248303, 'name': PowerOffVM_Task, 'duration_secs': 0.195775} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.918052] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.918267] env[63175]: DEBUG nova.compute.manager [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.919028] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb50a123-0807-4acf-9832-d34a45c2fa87 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.080555] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcfeb19-7f44-471c-b9b3-8da84133c518 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.088980] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fa84c6-d923-4a2a-9cfe-638153ba7669 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.119349] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15f8a9e-7c9a-4d2e-977d-e8a244125049 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.126855] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b33ffc4-c59e-4093-ac6d-4b670f8598c9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.139729] env[63175]: DEBUG nova.compute.provider_tree [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.159564] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.159670] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Instance network_info: |[{"id": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "address": "fa:16:3e:b1:53:28", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaec69ba-7e", "ovs_interfaceid": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 900.161031] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:53:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '721c6720-3ce0-450e-9951-a894f03acc27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eaec69ba-7e1a-4abf-b799-33836e2e6c04', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.167933] env[63175]: DEBUG oslo.service.loopingcall [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.168952] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.168952] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-709d004b-6631-485f-9658-7f926c201e67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.189142] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.189142] env[63175]: value = "task-1248304" [ 900.189142] env[63175]: _type = "Task" [ 900.189142] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.197535] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248304, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.262450] env[63175]: DEBUG nova.objects.instance [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.430992] env[63175]: DEBUG nova.compute.manager [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Received event network-changed-eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 900.433819] env[63175]: DEBUG nova.compute.manager [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Refreshing instance network info cache due to event network-changed-eaec69ba-7e1a-4abf-b799-33836e2e6c04. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 900.433819] env[63175]: DEBUG oslo_concurrency.lockutils [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] Acquiring lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.433819] env[63175]: DEBUG oslo_concurrency.lockutils [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] Acquired lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.433819] env[63175]: DEBUG nova.network.neutron [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Refreshing network info cache for port eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.440883] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a501bcfd-57f4-4887-89f2-4ebf3a6cb0a6 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.061s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.644875] env[63175]: DEBUG nova.scheduler.client.report [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 900.699387] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248304, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.825152] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.825445] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.825670] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.825857] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.826048] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.828299] env[63175]: INFO nova.compute.manager [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Terminating instance [ 901.134260] env[63175]: DEBUG nova.network.neutron [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Updated VIF entry in instance network info cache for port eaec69ba-7e1a-4abf-b799-33836e2e6c04. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.134714] env[63175]: DEBUG nova.network.neutron [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Updating instance_info_cache with network_info: [{"id": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "address": "fa:16:3e:b1:53:28", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaec69ba-7e", "ovs_interfaceid": "eaec69ba-7e1a-4abf-b799-33836e2e6c04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.150554] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.240s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.153093] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.300s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.154678] env[63175]: INFO nova.compute.claims [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.186818] env[63175]: INFO nova.network.neutron [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating port 6a69f7df-439f-4044-8394-305f9f3a93bb with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 901.199630] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248304, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.207459] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.269853] env[63175]: DEBUG oslo_concurrency.lockutils [None req-be304903-f6e2-4596-ba6c-d0670448967a tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.270987] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.064s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.271212] env[63175]: DEBUG nova.compute.manager [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 901.272293] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed8b5be-5cd4-4ef7-abaf-805bc91ab58f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.280156] env[63175]: DEBUG nova.compute.manager [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63175) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 901.280752] env[63175]: DEBUG nova.objects.instance [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.332051] env[63175]: DEBUG nova.compute.manager [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 901.332714] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.333535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccad5cd-2239-4916-9c97-26fbc90d5700 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.340991] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.341275] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c4cd1fe-8c8a-4ccf-8702-4cee14f450d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.418992] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.419301] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.419512] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleting the datastore file [datastore2] 8d3bd040-fa87-44b8-a6c1-fe6be21b0415 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.419784] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b968a33a-694c-402e-a703-ff5ad2c0a8ed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.427981] env[63175]: DEBUG oslo_vmware.api [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 901.427981] env[63175]: value = "task-1248306" [ 901.427981] env[63175]: _type = "Task" [ 901.427981] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.436564] env[63175]: DEBUG oslo_vmware.api [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.638357] env[63175]: DEBUG oslo_concurrency.lockutils [req-46c0f1a3-db3e-4b59-a3a2-d9d03db1613f req-983c89e4-3a34-429d-bb63-309fe9718f9a service nova] Releasing lock "refresh_cache-43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.699664] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248304, 'name': CreateVM_Task, 'duration_secs': 1.363121} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.699826] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.700550] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.700686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.701010] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.701508] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd183b5c-61f5-4832-a081-6715785f15df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.706138] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 901.706138] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ec685-c840-6d71-cd57-e2f764dcfdee" [ 901.706138] env[63175]: _type = "Task" [ 901.706138] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.713660] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ec685-c840-6d71-cd57-e2f764dcfdee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.940554] env[63175]: DEBUG oslo_vmware.api [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164654} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.940826] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.941032] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.941228] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.941407] env[63175]: INFO nova.compute.manager [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Took 0.61 seconds to destroy the instance on the hypervisor. [ 901.941683] env[63175]: DEBUG oslo.service.loopingcall [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.941890] env[63175]: DEBUG nova.compute.manager [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 901.941985] env[63175]: DEBUG nova.network.neutron [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 902.217038] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522ec685-c840-6d71-cd57-e2f764dcfdee, 'name': SearchDatastore_Task, 'duration_secs': 0.008929} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.219871] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.220135] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.220380] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.220733] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.220978] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.221707] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17a5ac30-ff22-4983-8d6e-09faa6bb238a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.230608] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.230804] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.233933] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53991820-fbba-48b2-9073-70c065d7b356 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.240419] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 902.240419] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52903677-aeba-0f1f-d45b-03353d0dabf5" [ 902.240419] env[63175]: _type = "Task" [ 902.240419] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.249088] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52903677-aeba-0f1f-d45b-03353d0dabf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.287944] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.288248] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af0bf1d3-d9fa-477c-a46d-4efae8d37668 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.296909] env[63175]: DEBUG oslo_vmware.api [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 902.296909] env[63175]: value = "task-1248307" [ 902.296909] env[63175]: _type = "Task" [ 902.296909] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.309504] env[63175]: DEBUG oslo_vmware.api [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248307, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.363719] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1b7725-d26b-40d7-8202-d19d0717b17b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.372674] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6027e7bf-1276-4390-9b46-40cb5d693fa0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.405424] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa675a5-7a2e-459b-9583-371d83e16366 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.414059] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fadbed-e904-43b4-898d-49ab48cd2d67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.431223] env[63175]: DEBUG nova.compute.provider_tree [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.463234] env[63175]: DEBUG nova.compute.manager [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 902.463557] env[63175]: DEBUG oslo_concurrency.lockutils [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.463709] env[63175]: DEBUG oslo_concurrency.lockutils [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.463888] env[63175]: DEBUG oslo_concurrency.lockutils [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.464072] env[63175]: DEBUG nova.compute.manager [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] No waiting events found dispatching network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 902.464243] env[63175]: WARNING nova.compute.manager [req-8e6ffe7d-54ae-4252-a7d7-b376f21498b4 req-1b81dc0e-16b7-4fdf-83c4-6c85c472ac8b service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received unexpected event network-vif-plugged-6a69f7df-439f-4044-8394-305f9f3a93bb for instance with vm_state shelved_offloaded and task_state spawning. [ 902.606742] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.606977] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.607388] env[63175]: DEBUG nova.network.neutron [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.751217] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52903677-aeba-0f1f-d45b-03353d0dabf5, 'name': SearchDatastore_Task, 'duration_secs': 0.009652} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.752077] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-127e5946-c2b7-4cf7-a973-92a144b8c390 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.757608] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 902.757608] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524ecb6e-e256-a129-0410-7ca261539371" [ 902.757608] env[63175]: _type = "Task" [ 902.757608] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.765162] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524ecb6e-e256-a129-0410-7ca261539371, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.806278] env[63175]: DEBUG oslo_vmware.api [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248307, 'name': PowerOffVM_Task, 'duration_secs': 0.196769} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.806553] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.806756] env[63175]: DEBUG nova.compute.manager [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.807614] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a5d2b7-ce65-435e-acb3-6d3f6ae18286 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.934579] env[63175]: DEBUG nova.scheduler.client.report [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 903.064923] env[63175]: DEBUG nova.network.neutron [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.269007] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524ecb6e-e256-a129-0410-7ca261539371, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.271631] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.271900] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8/43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.272176] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffb38e2a-e8f4-4de1-9e51-43c93ae4af3f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.280088] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 903.280088] env[63175]: value = "task-1248308" [ 903.280088] env[63175]: _type = "Task" [ 903.280088] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.290144] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.318899] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4681648a-701e-4422-9f13-3fe7f94ec31c tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.048s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.346696] env[63175]: DEBUG nova.network.neutron [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.439164] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.439686] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 903.442979] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.559s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.443219] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.476051] env[63175]: INFO nova.scheduler.client.report [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocations for instance 4bd91412-c41b-41a1-a648-6b905d826ee3 [ 903.567033] env[63175]: INFO nova.compute.manager [-] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Took 1.62 seconds to deallocate network for instance. [ 903.792435] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506664} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.792712] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8/43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.792935] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.793215] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c5010bd-1f57-4b6b-ad71-92881926c037 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.800699] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 903.800699] env[63175]: value = "task-1248309" [ 903.800699] env[63175]: _type = "Task" [ 903.800699] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.810100] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248309, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.849242] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.877272] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='cf237f8a70f1aa704a80a58f8a82e26e',container_format='bare',created_at=2024-10-10T14:50:30Z,direct_url=,disk_format='vmdk',id=d8860fd4-87bf-46ae-93cb-6b6f2f8627cd,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-587117315-shelved',owner='5ea6da3858c347d887dcfba2556fabad',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-10T14:50:44Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.877515] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.877710] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.877889] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.878089] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.878266] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.878488] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.878678] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.878884] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.879100] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.879294] env[63175]: DEBUG nova.virt.hardware [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.880162] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95242e0-c29c-4216-9e37-3394f8edb661 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.888850] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfa83b1-7779-42d4-b8d8-7566ab5db26a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.902810] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:c0:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a69f7df-439f-4044-8394-305f9f3a93bb', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.910159] env[63175]: DEBUG oslo.service.loopingcall [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.910420] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.910632] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a86175b-f804-4a4a-9730-27684fd6d4ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.930182] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.930182] env[63175]: value = "task-1248310" [ 903.930182] env[63175]: _type = "Task" [ 903.930182] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.934547] env[63175]: DEBUG nova.objects.instance [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.942057] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248310, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.944480] env[63175]: DEBUG nova.compute.utils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.945820] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 903.945994] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.984555] env[63175]: DEBUG oslo_concurrency.lockutils [None req-69d37a50-bb39-429f-aaf9-7b7c5bcd915b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "4bd91412-c41b-41a1-a648-6b905d826ee3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.591s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.998411] env[63175]: DEBUG nova.policy [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c654f31fd4bc49dc8e985bd75a811ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74dd39e20aad4ddb874282973cea02bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.074320] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.074639] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.075073] env[63175]: DEBUG nova.objects.instance [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'resources' on Instance uuid 8d3bd040-fa87-44b8-a6c1-fe6be21b0415 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.262646] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Successfully created port: 95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.310393] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248309, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065724} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.310661] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.311465] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb51fa3-54a7-46d2-9288-38ebb9a21f13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.333691] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8/43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.333983] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acc1c7d8-e5d6-424f-8d51-40218aec4fda {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.354333] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 904.354333] env[63175]: value = "task-1248311" [ 904.354333] env[63175]: _type = "Task" [ 904.354333] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.362531] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.441561] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248310, 'name': CreateVM_Task, 'duration_secs': 0.341029} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.442095] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.442272] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.442457] env[63175]: DEBUG nova.network.neutron [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.442679] env[63175]: DEBUG nova.objects.instance [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'info_cache' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.443776] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.444606] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.444786] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.445208] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.445501] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e141a290-af0e-40bf-b9a7-abe3bd2f6518 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.450792] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 904.453599] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 904.453599] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d64c8a-7c20-8395-9d98-92faea05467c" [ 904.453599] env[63175]: _type = "Task" [ 904.453599] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.465183] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d64c8a-7c20-8395-9d98-92faea05467c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.496127] env[63175]: DEBUG nova.compute.manager [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Received event network-vif-deleted-8ffc5202-b3e9-45c5-9821-2b18e73a48e8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 904.497302] env[63175]: DEBUG nova.compute.manager [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 904.497580] env[63175]: DEBUG nova.compute.manager [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing instance network info cache due to event network-changed-6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 904.497769] env[63175]: DEBUG oslo_concurrency.lockutils [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] Acquiring lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.497865] env[63175]: DEBUG oslo_concurrency.lockutils [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] Acquired lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.498075] env[63175]: DEBUG nova.network.neutron [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Refreshing network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.714479] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c0617b-9483-4352-9c8c-9f19f8dda530 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.723268] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80370025-ef07-49ce-8ad0-3b5eee3a60a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.753589] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05bbc1a-253f-46a5-b1bb-e23124fd2db3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.762270] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6fccc1-f905-43b4-a4ff-e8a4dfc2bb8f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.776590] env[63175]: DEBUG nova.compute.provider_tree [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.865666] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248311, 'name': ReconfigVM_Task, 'duration_secs': 0.272197} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.866015] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8/43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.866653] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cfdf3f8f-ed30-4cbe-ac0f-7363b3847774 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.873764] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 904.873764] env[63175]: value = "task-1248312" [ 904.873764] env[63175]: _type = "Task" [ 904.873764] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.884309] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248312, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.945565] env[63175]: DEBUG nova.objects.base [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.967755] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.968020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Processing image d8860fd4-87bf-46ae-93cb-6b6f2f8627cd {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.968257] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.968407] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.968593] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.968838] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c450010d-f56f-4d0e-a052-82b535a0a7d3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.989538] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.989807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.991181] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a7054c-2fe5-4256-bad9-3a86e437e185 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.998033] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 904.998033] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163b34-6fd9-89c8-579e-f278adc9f8df" [ 904.998033] env[63175]: _type = "Task" [ 904.998033] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.008943] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52163b34-6fd9-89c8-579e-f278adc9f8df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.279883] env[63175]: DEBUG nova.scheduler.client.report [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 905.296033] env[63175]: DEBUG nova.network.neutron [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updated VIF entry in instance network info cache for port 6a69f7df-439f-4044-8394-305f9f3a93bb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.296131] env[63175]: DEBUG nova.network.neutron [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [{"id": "6a69f7df-439f-4044-8394-305f9f3a93bb", "address": "fa:16:3e:94:c0:d2", "network": {"id": "1ca32f9a-6af3-4e8b-b41b-807d806795e7", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-752815788-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea6da3858c347d887dcfba2556fabad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a69f7df-43", "ovs_interfaceid": "6a69f7df-439f-4044-8394-305f9f3a93bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.384131] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248312, 'name': Rename_Task, 'duration_secs': 0.144124} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.384415] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.384643] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-239c278b-8603-4a0e-af69-30bebd54fbd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.391075] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 905.391075] env[63175]: value = "task-1248313" [ 905.391075] env[63175]: _type = "Task" [ 905.391075] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.398717] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248313, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.462324] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 905.488499] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.488694] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.488858] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.489066] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.489227] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.489377] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.489593] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.489755] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.489924] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.490104] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.490282] env[63175]: DEBUG nova.virt.hardware [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.491156] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4ec212-f8aa-4d25-ac51-c46a8471fff0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.502656] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3b2c2f-1ebb-45a9-bb1f-b246bcdb9301 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.520626] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Preparing fetch location {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 905.520891] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Fetch image to [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0/OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0.vmdk {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 905.521095] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Downloading stream optimized image d8860fd4-87bf-46ae-93cb-6b6f2f8627cd to [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0/OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0.vmdk on the data store datastore2 as vApp {{(pid=63175) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 905.521326] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Downloading image file data d8860fd4-87bf-46ae-93cb-6b6f2f8627cd to the ESX as VM named 'OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0' {{(pid=63175) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 905.595381] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 905.595381] env[63175]: value = "resgroup-9" [ 905.595381] env[63175]: _type = "ResourcePool" [ 905.595381] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 905.595760] env[63175]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-aa927342-8373-492b-b834-44efb6c1106a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.619630] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lease: (returnval){ [ 905.619630] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 905.619630] env[63175]: _type = "HttpNfcLease" [ 905.619630] env[63175]: } obtained for vApp import into resource pool (val){ [ 905.619630] env[63175]: value = "resgroup-9" [ 905.619630] env[63175]: _type = "ResourcePool" [ 905.619630] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 905.619918] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the lease: (returnval){ [ 905.619918] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 905.619918] env[63175]: _type = "HttpNfcLease" [ 905.619918] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 905.626771] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.626771] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 905.626771] env[63175]: _type = "HttpNfcLease" [ 905.626771] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 905.784537] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.799703] env[63175]: DEBUG oslo_concurrency.lockutils [req-1959fde6-c71b-4d32-9c79-5f4b5e30a21a req-c8e8b7a3-0e34-4ed0-84b3-1a9f26e6b8fb service nova] Releasing lock "refresh_cache-5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.809345] env[63175]: INFO nova.scheduler.client.report [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance 8d3bd040-fa87-44b8-a6c1-fe6be21b0415 [ 905.833198] env[63175]: DEBUG nova.network.neutron [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.902921] env[63175]: DEBUG oslo_vmware.api [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248313, 'name': PowerOnVM_Task, 'duration_secs': 0.438065} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.902921] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.902921] env[63175]: INFO nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Took 7.79 seconds to spawn the instance on the hypervisor. [ 905.902921] env[63175]: DEBUG nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 905.902921] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c928881-0a03-4866-ae3c-8009a6a54206 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.979841] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Successfully updated port: 95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.127639] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 906.127639] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 906.127639] env[63175]: _type = "HttpNfcLease" [ 906.127639] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 906.318825] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88e1235d-0bae-4671-95cd-ae096b03f1ba tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "8d3bd040-fa87-44b8-a6c1-fe6be21b0415" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.493s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.338256] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.420547] env[63175]: INFO nova.compute.manager [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Took 16.68 seconds to build instance. [ 906.483064] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.483262] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.483454] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.521706] env[63175]: DEBUG nova.compute.manager [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Received event network-vif-plugged-95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 906.521706] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.522082] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.522082] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.522153] env[63175]: DEBUG nova.compute.manager [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] No waiting events found dispatching network-vif-plugged-95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 906.523181] env[63175]: WARNING nova.compute.manager [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Received unexpected event network-vif-plugged-95854088-699b-4fec-a714-1b6959135fd2 for instance with vm_state building and task_state spawning. [ 906.523181] env[63175]: DEBUG nova.compute.manager [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Received event network-changed-95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 906.523181] env[63175]: DEBUG nova.compute.manager [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Refreshing instance network info cache due to event network-changed-95854088-699b-4fec-a714-1b6959135fd2. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 906.523181] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Acquiring lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.629450] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 906.629450] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 906.629450] env[63175]: _type = "HttpNfcLease" [ 906.629450] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 906.629761] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 906.629761] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3666e-dcff-480d-73cf-4388f252cd5a" [ 906.629761] env[63175]: _type = "HttpNfcLease" [ 906.629761] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 906.630431] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8354c21a-8888-4992-81be-d144ae5004cc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.637671] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 906.637873] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk. {{(pid=63175) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 906.700941] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-15a042eb-0ddc-4ace-99dc-142021f64b9d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.923518] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8dc52032-e663-494e-8a64-97583b87a89f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.198s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.033070] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.259715] env[63175]: DEBUG nova.network.neutron [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.344039] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.344289] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7147170d-cf46-4e02-951a-90b9a8724c56 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.353616] env[63175]: DEBUG oslo_vmware.api [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 907.353616] env[63175]: value = "task-1248315" [ 907.353616] env[63175]: _type = "Task" [ 907.353616] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.364904] env[63175]: DEBUG oslo_vmware.api [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248315, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.661085] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb4483c-3ef6-4c1f-a29e-e88ce444a550 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.671168] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Suspending the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 907.671535] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-461b503e-ae50-4bdf-a1cf-27f6e621c0ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.680316] env[63175]: DEBUG oslo_vmware.api [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 907.680316] env[63175]: value = "task-1248316" [ 907.680316] env[63175]: _type = "Task" [ 907.680316] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.691649] env[63175]: DEBUG oslo_vmware.api [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248316, 'name': SuspendVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.763160] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.763693] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Instance network_info: |[{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 907.764078] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Acquired lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.764284] env[63175]: DEBUG nova.network.neutron [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Refreshing network info cache for port 95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.765618] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:d9:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95854088-699b-4fec-a714-1b6959135fd2', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.773797] env[63175]: DEBUG oslo.service.loopingcall [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.774396] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.774676] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52bbf86c-4113-4b02-b5a9-3d4d792fa7fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.804410] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.804410] env[63175]: value = "task-1248317" [ 907.804410] env[63175]: _type = "Task" [ 907.804410] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.815984] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248317, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.866902] env[63175]: DEBUG oslo_vmware.api [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248315, 'name': PowerOnVM_Task, 'duration_secs': 0.484064} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.867212] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.867436] env[63175]: DEBUG nova.compute.manager [None req-0fa2e66e-f74e-4eea-8a9e-c019d920c0e8 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 907.868647] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297acc2c-f866-4fcf-9446-e62e7df569f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.883974] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Completed reading data from the image iterator. {{(pid=63175) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 907.884397] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 907.886093] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17640e71-6052-43c1-8f98-e885ba03e04e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.894424] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 907.894785] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 907.895050] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-881b2154-5573-4249-a607-72fb0b528a60 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.016000] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "ed618dba-1b54-4119-a317-4f64a64a59c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.016000] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.192595] env[63175]: DEBUG oslo_vmware.api [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248316, 'name': SuspendVM_Task} progress is 62%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.232231] env[63175]: DEBUG oslo_vmware.rw_handles [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526c2465-8d08-a74b-fb29-f8e03a444dcc/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 908.232470] env[63175]: INFO nova.virt.vmwareapi.images [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Downloaded image file data d8860fd4-87bf-46ae-93cb-6b6f2f8627cd [ 908.233319] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2ea5e1-0cc8-43a7-a409-3e04d5dd2dc3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.249907] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f94d4526-431c-4f70-aacb-355ac44c1690 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.288202] env[63175]: INFO nova.virt.vmwareapi.images [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] The imported VM was unregistered [ 908.291367] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Caching image {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 908.291901] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.292304] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-454849a4-c1bb-4ea0-9f21-ceab99649298 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.309678] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Created directory with path [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.309958] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0/OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0.vmdk to [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk. {{(pid=63175) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 908.310618] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d507ca2c-e71b-46f6-8a0c-50fc1d03b8ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.319136] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248317, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.324237] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 908.324237] env[63175]: value = "task-1248319" [ 908.324237] env[63175]: _type = "Task" [ 908.324237] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.332879] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.506045] env[63175]: DEBUG nova.network.neutron [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updated VIF entry in instance network info cache for port 95854088-699b-4fec-a714-1b6959135fd2. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.506454] env[63175]: DEBUG nova.network.neutron [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.521072] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 908.692798] env[63175]: DEBUG oslo_vmware.api [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248316, 'name': SuspendVM_Task, 'duration_secs': 0.659396} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.693150] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Suspended the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 908.693253] env[63175]: DEBUG nova.compute.manager [None req-c63e10c8-92b4-49c6-823d-5855a4688366 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.693991] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86246033-8963-425c-a34d-04900fa1d13b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.815593] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248317, 'name': CreateVM_Task, 'duration_secs': 0.56848} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.815797] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.816578] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.816808] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.817231] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.817533] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-107e711b-940f-4b79-be62-0b1202893c73 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.822410] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 908.822410] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e26f41-e9bd-ba02-b021-4f717bed49ae" [ 908.822410] env[63175]: _type = "Task" [ 908.822410] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.833084] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.835868] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e26f41-e9bd-ba02-b021-4f717bed49ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.009457] env[63175]: DEBUG oslo_concurrency.lockutils [req-f0511734-f333-405d-87db-9905dbb6d5b9 req-ee37490e-6247-4785-9471-acfafdcf65b4 service nova] Releasing lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.046872] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.047067] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.048684] env[63175]: INFO nova.compute.claims [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.336779] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.340534] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e26f41-e9bd-ba02-b021-4f717bed49ae, 'name': SearchDatastore_Task, 'duration_secs': 0.044289} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.340833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.341085] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.341352] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.341523] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.341712] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.341989] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16d51d77-19df-44a9-97b9-3d03cc4c824d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.360951] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.361186] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.362098] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb446cfc-16de-4e83-bb27-dbf3f06a8ec7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.370965] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 909.370965] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523a646d-abdb-dad7-0f02-d9d36a13c05a" [ 909.370965] env[63175]: _type = "Task" [ 909.370965] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.380965] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523a646d-abdb-dad7-0f02-d9d36a13c05a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.840618] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.885255] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523a646d-abdb-dad7-0f02-d9d36a13c05a, 'name': SearchDatastore_Task, 'duration_secs': 0.078316} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.886099] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c030676-ea54-4dcf-8136-74a4ce793068 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.895231] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 909.895231] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52395a04-dfa0-303c-17a9-f704de4b00d5" [ 909.895231] env[63175]: _type = "Task" [ 909.895231] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.907340] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52395a04-dfa0-303c-17a9-f704de4b00d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.217753] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea0a840-7362-4611-90b6-51d7a8f54b90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.228662] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336bcbf3-a8e0-4f43-bc48-ea5c8d8ab7dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.264730] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6d066b-c919-4c9d-aa45-813e460923d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.275609] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53057fa9-4e13-46c4-be05-008ba7f79003 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.292579] env[63175]: DEBUG nova.compute.provider_tree [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.342810] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.408154] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52395a04-dfa0-303c-17a9-f704de4b00d5, 'name': SearchDatastore_Task, 'duration_secs': 0.082462} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.408460] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.408740] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.409035] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac444829-314d-4792-a654-f2d3e1dedcc3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.420566] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 910.420566] env[63175]: value = "task-1248320" [ 910.420566] env[63175]: _type = "Task" [ 910.420566] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.434174] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.795043] env[63175]: DEBUG nova.scheduler.client.report [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 910.841472] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.933376] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.978084] env[63175]: DEBUG nova.compute.manager [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 910.978406] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5012d636-f2e1-4c86-9aeb-162d0421d1f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.300587] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.301094] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 911.344245] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.430733] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.491714] env[63175]: INFO nova.compute.manager [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] instance snapshotting [ 911.491986] env[63175]: WARNING nova.compute.manager [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 911.495114] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4183a1-78c7-4cd6-a474-7a34aebe74f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.515053] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4957224a-74ea-44c2-99e2-1b211f110557 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.805904] env[63175]: DEBUG nova.compute.utils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.807399] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 911.807573] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.840748] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248319, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.448337} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.841480] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0/OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0.vmdk to [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk. [ 911.841480] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Cleaning up location [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 911.841827] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_df1280c4-1d23-4612-89b1-3e178d86c0d0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.841868] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1eb98fac-d7a9-4ede-8a29-babad3758742 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.847345] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 911.847345] env[63175]: value = "task-1248321" [ 911.847345] env[63175]: _type = "Task" [ 911.847345] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.848636] env[63175]: DEBUG nova.policy [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.857528] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.930376] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248320, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.026178] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 912.026511] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3e97999d-476e-451b-b3ee-44db493e45b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.034247] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 912.034247] env[63175]: value = "task-1248322" [ 912.034247] env[63175]: _type = "Task" [ 912.034247] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.044861] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248322, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.177045] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Successfully created port: 4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.311017] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 912.359768] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153964} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.360050] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.360226] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.360474] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk to [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.360724] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8e2877d-192f-458c-a8c7-0257e551fb4e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.366971] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 912.366971] env[63175]: value = "task-1248323" [ 912.366971] env[63175]: _type = "Task" [ 912.366971] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.374929] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.431550] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248320, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.920219} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.431550] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.431759] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.431960] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47230700-9ae6-4e1b-88ee-02d5d30bcd91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.438516] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 912.438516] env[63175]: value = "task-1248324" [ 912.438516] env[63175]: _type = "Task" [ 912.438516] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.446626] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.544586] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248322, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.881531] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.951269] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071018} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.952015] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.952687] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53c2070-c3f0-46f8-a975-3e0c55758df2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.977461] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.977853] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-338769cd-6f0a-43a9-909c-576a26b09747 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.001345] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 913.001345] env[63175]: value = "task-1248325" [ 913.001345] env[63175]: _type = "Task" [ 913.001345] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.012447] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.046801] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248322, 'name': CreateSnapshot_Task, 'duration_secs': 0.571685} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.047095] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 913.047869] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03ed91c-11a6-450c-8fab-94d0fefd72ac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.323821] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 913.349819] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.350147] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.350346] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.350611] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.351679] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.351679] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.351679] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.351679] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.351856] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.352063] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.352268] env[63175]: DEBUG nova.virt.hardware [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.353706] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514b795b-f737-46c7-8914-934850d18859 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.365399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec20b23-fb95-48ed-b560-7b47960a7dbc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.376897] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.512783] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.568712] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 913.569145] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7384c36a-eb15-4bcb-8868-9847a480d83a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.579601] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 913.579601] env[63175]: value = "task-1248326" [ 913.579601] env[63175]: _type = "Task" [ 913.579601] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.590074] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248326, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.608054] env[63175]: DEBUG nova.compute.manager [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Received event network-vif-plugged-4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 913.608296] env[63175]: DEBUG oslo_concurrency.lockutils [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] Acquiring lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.608512] env[63175]: DEBUG oslo_concurrency.lockutils [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.608688] env[63175]: DEBUG oslo_concurrency.lockutils [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.608865] env[63175]: DEBUG nova.compute.manager [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] No waiting events found dispatching network-vif-plugged-4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 913.609045] env[63175]: WARNING nova.compute.manager [req-21dea9a6-cc48-495b-b2a6-3e7cc1d4e821 req-95f474e9-ed7a-4ff4-85d5-00ab644cc6e5 service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Received unexpected event network-vif-plugged-4183e5b2-8de0-4874-a2b8-6a15994712b5 for instance with vm_state building and task_state spawning. [ 913.721450] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Successfully updated port: 4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.882011] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.014441] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.092654] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248326, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.226356] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.226513] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.226725] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.381698] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.514738] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.592455] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248326, 'name': CloneVM_Task} progress is 95%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.759747] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.879719] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248323, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.42239} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.879999] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd/d8860fd4-87bf-46ae-93cb-6b6f2f8627cd.vmdk to [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 914.880788] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6e49b4-a7a1-48cd-a2c2-286974ab7e67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.905203] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.906295] env[63175]: DEBUG nova.network.neutron [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Updating instance_info_cache with network_info: [{"id": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "address": "fa:16:3e:6b:ba:3b", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4183e5b2-8d", "ovs_interfaceid": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.907453] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e7d5162-b831-4f65-9aaf-adab96915f22 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.922924] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.923264] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Instance network_info: |[{"id": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "address": "fa:16:3e:6b:ba:3b", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4183e5b2-8d", "ovs_interfaceid": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 914.924373] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:ba:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4183e5b2-8de0-4874-a2b8-6a15994712b5', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.931674] env[63175]: DEBUG oslo.service.loopingcall [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.932938] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.933201] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7f8fb4f-0a02-4a62-b106-2ab19c50e8fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.949248] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 914.949248] env[63175]: value = "task-1248327" [ 914.949248] env[63175]: _type = "Task" [ 914.949248] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.957892] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248327, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.959018] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.959018] env[63175]: value = "task-1248328" [ 914.959018] env[63175]: _type = "Task" [ 914.959018] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.967639] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248328, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.012933] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.091069] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248326, 'name': CloneVM_Task, 'duration_secs': 1.34033} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.091365] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Created linked-clone VM from snapshot [ 915.092119] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9eddf2a-e03b-4b0e-947e-a05848f984a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.100168] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Uploading image ea8ff57a-aeaf-4d9b-878b-d9b0851306bf {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 915.120944] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 915.120944] env[63175]: value = "vm-269068" [ 915.120944] env[63175]: _type = "VirtualMachine" [ 915.120944] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 915.121249] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-de48497c-00c1-4251-9c2f-6e853c9003c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.129139] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease: (returnval){ [ 915.129139] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dbb35e-2e57-6572-d8ec-094ab974c449" [ 915.129139] env[63175]: _type = "HttpNfcLease" [ 915.129139] env[63175]: } obtained for exporting VM: (result){ [ 915.129139] env[63175]: value = "vm-269068" [ 915.129139] env[63175]: _type = "VirtualMachine" [ 915.129139] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 915.129477] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the lease: (returnval){ [ 915.129477] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dbb35e-2e57-6572-d8ec-094ab974c449" [ 915.129477] env[63175]: _type = "HttpNfcLease" [ 915.129477] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 915.135732] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.135732] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dbb35e-2e57-6572-d8ec-094ab974c449" [ 915.135732] env[63175]: _type = "HttpNfcLease" [ 915.135732] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 915.459114] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248327, 'name': ReconfigVM_Task, 'duration_secs': 0.306602} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.459417] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f/5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.460096] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74cb0e8d-fe68-495f-b6ef-56f6bac72935 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.469362] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248328, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.470549] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 915.470549] env[63175]: value = "task-1248330" [ 915.470549] env[63175]: _type = "Task" [ 915.470549] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.477517] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248330, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.513263] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248325, 'name': ReconfigVM_Task, 'duration_secs': 2.14677} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.513685] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfigured VM instance instance-00000051 to attach disk [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.514367] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cfaa9f1-2ca3-41db-a567-776b6cc9e09c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.520251] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 915.520251] env[63175]: value = "task-1248331" [ 915.520251] env[63175]: _type = "Task" [ 915.520251] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.528551] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248331, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.635517] env[63175]: DEBUG nova.compute.manager [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Received event network-changed-4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 915.635852] env[63175]: DEBUG nova.compute.manager [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Refreshing instance network info cache due to event network-changed-4183e5b2-8de0-4874-a2b8-6a15994712b5. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 915.636019] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] Acquiring lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.636204] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] Acquired lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.636377] env[63175]: DEBUG nova.network.neutron [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Refreshing network info cache for port 4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.642077] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.642077] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dbb35e-2e57-6572-d8ec-094ab974c449" [ 915.642077] env[63175]: _type = "HttpNfcLease" [ 915.642077] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 915.642368] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 915.642368] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dbb35e-2e57-6572-d8ec-094ab974c449" [ 915.642368] env[63175]: _type = "HttpNfcLease" [ 915.642368] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 915.643155] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d698aa-0657-4275-b185-9076f182119f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.652332] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 915.652524] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 915.740032] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7dee17f5-da17-4329-a4b8-623547e8ef06 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.972150] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248328, 'name': CreateVM_Task, 'duration_secs': 0.588348} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.975239] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.975989] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.976220] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.976558] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.977183] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-131b902b-1626-47bd-9df7-bd3e2e71ddde {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.982297] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248330, 'name': Rename_Task, 'duration_secs': 0.157591} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.983593] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.983955] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 915.983955] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a9e5b8-bb45-2eba-a602-dcf8a37cc3a3" [ 915.983955] env[63175]: _type = "Task" [ 915.983955] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.984156] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb4039a3-07ca-459b-97db-ec78a67d3311 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.995961] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a9e5b8-bb45-2eba-a602-dcf8a37cc3a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.997419] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 915.997419] env[63175]: value = "task-1248332" [ 915.997419] env[63175]: _type = "Task" [ 915.997419] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.005489] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248332, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.031552] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248331, 'name': Rename_Task, 'duration_secs': 0.197549} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.031855] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.032167] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3aa3eee7-6b5d-438c-bac4-f8a955d3f222 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.038964] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 916.038964] env[63175]: value = "task-1248333" [ 916.038964] env[63175]: _type = "Task" [ 916.038964] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.049358] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248333, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.358792] env[63175]: DEBUG nova.network.neutron [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Updated VIF entry in instance network info cache for port 4183e5b2-8de0-4874-a2b8-6a15994712b5. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.359319] env[63175]: DEBUG nova.network.neutron [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Updating instance_info_cache with network_info: [{"id": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "address": "fa:16:3e:6b:ba:3b", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4183e5b2-8d", "ovs_interfaceid": "4183e5b2-8de0-4874-a2b8-6a15994712b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.503273] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a9e5b8-bb45-2eba-a602-dcf8a37cc3a3, 'name': SearchDatastore_Task, 'duration_secs': 0.014199} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.509311] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.510043] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.510779] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.511188] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.511484] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.511922] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93a54f2f-fcc2-4675-befe-4213b8b125ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.523705] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248332, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.536708] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.537017] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.537744] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-256c01a3-4b91-477f-b418-32cd3ae6a7a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.549599] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 916.549599] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5282f93b-739d-efb8-d24e-12f7060c3c6a" [ 916.549599] env[63175]: _type = "Task" [ 916.549599] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.554189] env[63175]: DEBUG oslo_vmware.api [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248333, 'name': PowerOnVM_Task, 'duration_secs': 0.492956} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.557875] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.558293] env[63175]: INFO nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Took 11.10 seconds to spawn the instance on the hypervisor. [ 916.558520] env[63175]: DEBUG nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.559393] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4157adb-628e-485a-8c38-907f04d8f799 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.569790] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5282f93b-739d-efb8-d24e-12f7060c3c6a, 'name': SearchDatastore_Task, 'duration_secs': 0.01426} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.575469] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83b137e5-598f-49a7-a086-a62988f837c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.582872] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 916.582872] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c7601b-9ca7-6b4d-a005-c1a3f8e489ec" [ 916.582872] env[63175]: _type = "Task" [ 916.582872] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.595429] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c7601b-9ca7-6b4d-a005-c1a3f8e489ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.863210] env[63175]: DEBUG oslo_concurrency.lockutils [req-fe8d9ba2-b0a7-4edd-8eeb-b4caed63b69c req-686bfe7a-447b-4e83-b353-cb1500824c5f service nova] Releasing lock "refresh_cache-ed618dba-1b54-4119-a317-4f64a64a59c9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.018408] env[63175]: DEBUG oslo_vmware.api [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248332, 'name': PowerOnVM_Task, 'duration_secs': 0.563185} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.018838] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.090927] env[63175]: INFO nova.compute.manager [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Took 25.25 seconds to build instance. [ 917.099131] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c7601b-9ca7-6b4d-a005-c1a3f8e489ec, 'name': SearchDatastore_Task, 'duration_secs': 0.019422} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.099642] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.100065] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] ed618dba-1b54-4119-a317-4f64a64a59c9/ed618dba-1b54-4119-a317-4f64a64a59c9.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.100450] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-337cd1ac-9074-4bf6-9ea7-fec7746c07be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.111000] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 917.111000] env[63175]: value = "task-1248334" [ 917.111000] env[63175]: _type = "Task" [ 917.111000] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.122045] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248334, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.148637] env[63175]: DEBUG nova.compute.manager [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 917.149098] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebf2a85-c160-4d2f-9133-ae52082a6f7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.594814] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8be2e5c5-f15b-4bc6-bfc8-d694c657dd7b tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.768s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.625032] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248334, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.673225] env[63175]: DEBUG oslo_concurrency.lockutils [None req-def8dc2d-0cd2-44f8-ae82-af1c3cd3b2af tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 26.969s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.123869] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248334, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625185} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.124234] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] ed618dba-1b54-4119-a317-4f64a64a59c9/ed618dba-1b54-4119-a317-4f64a64a59c9.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.124418] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.124720] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af6a031b-0e99-4f12-a3b8-e531a1636e70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.135086] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 918.135086] env[63175]: value = "task-1248335" [ 918.135086] env[63175]: _type = "Task" [ 918.135086] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.143887] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248335, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.645966] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248335, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134601} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.646217] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.647192] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e31a96d-ca69-4a72-9054-a4b0c6c51402 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.669810] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] ed618dba-1b54-4119-a317-4f64a64a59c9/ed618dba-1b54-4119-a317-4f64a64a59c9.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.670138] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea79c2d8-072f-4c51-a755-5307cb590253 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.690593] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 918.690593] env[63175]: value = "task-1248336" [ 918.690593] env[63175]: _type = "Task" [ 918.690593] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.699091] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248336, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.201538] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248336, 'name': ReconfigVM_Task, 'duration_secs': 0.503388} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.201972] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Reconfigured VM instance instance-00000052 to attach disk [datastore1] ed618dba-1b54-4119-a317-4f64a64a59c9/ed618dba-1b54-4119-a317-4f64a64a59c9.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.202526] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a78dbc61-cc1e-4a6b-9a44-bcac2778e0fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.210087] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 919.210087] env[63175]: value = "task-1248337" [ 919.210087] env[63175]: _type = "Task" [ 919.210087] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.221207] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248337, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.319847] env[63175]: DEBUG nova.compute.manager [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Stashing vm_state: active {{(pid=63175) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 919.721064] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248337, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.840692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.840978] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.221358] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248337, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.347856] env[63175]: INFO nova.compute.claims [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.722222] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248337, 'name': Rename_Task, 'duration_secs': 1.149135} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.722549] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.722831] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71622faa-6efb-4750-8c1f-4cf6daed4023 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.730909] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 920.730909] env[63175]: value = "task-1248338" [ 920.730909] env[63175]: _type = "Task" [ 920.730909] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.739522] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.853055] env[63175]: INFO nova.compute.resource_tracker [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating resource usage from migration 757bbb05-bc15-40f2-b008-b8179561f844 [ 921.027780] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e5c66d-f36d-4de1-a676-41d47b9e0a4c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.037076] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80fea90-715f-43f7-8102-b784b326a111 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.069282] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a65a1f-4546-47bf-a7a8-d2b85be97665 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.078255] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdf2648-ef8a-4de5-af1d-5606a314e910 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.094487] env[63175]: DEBUG nova.compute.provider_tree [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.244282] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248338, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.597631] env[63175]: DEBUG nova.scheduler.client.report [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 921.742634] env[63175]: DEBUG oslo_vmware.api [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248338, 'name': PowerOnVM_Task, 'duration_secs': 0.722006} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.744016] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.744016] env[63175]: INFO nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Took 8.42 seconds to spawn the instance on the hypervisor. [ 921.744016] env[63175]: DEBUG nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 921.744793] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aceaef0-b160-46d1-82df-f912d8bceb74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.102537] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.261s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.102753] env[63175]: INFO nova.compute.manager [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Migrating [ 922.262190] env[63175]: INFO nova.compute.manager [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Took 13.24 seconds to build instance. [ 922.617720] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.618023] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.618995] env[63175]: DEBUG nova.network.neutron [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.737852] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 922.739017] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2dceab1-bd16-4e1e-ad62-05f8373a0da0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.746683] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 922.746859] env[63175]: ERROR oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk due to incomplete transfer. [ 922.747096] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-316b37de-7ed1-4e00-9c29-f40df707a179 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.753964] env[63175]: DEBUG oslo_vmware.rw_handles [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52569f92-af5d-7698-6c5e-95398ede82b8/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 922.754829] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Uploaded image ea8ff57a-aeaf-4d9b-878b-d9b0851306bf to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 922.756749] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 922.756749] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d0b3d188-2b31-4f30-929f-27265e25d658 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.763342] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 922.763342] env[63175]: value = "task-1248339" [ 922.763342] env[63175]: _type = "Task" [ 922.763342] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.763738] env[63175]: DEBUG oslo_concurrency.lockutils [None req-565c5a98-c8bb-4ffd-92a9-6ab2754e532a tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.748s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.771958] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248339, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.273175] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248339, 'name': Destroy_Task, 'duration_secs': 0.330169} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.273502] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Destroyed the VM [ 923.273706] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 923.273959] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-95f39282-f6b0-4bee-95d9-f31d746f56dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.281369] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 923.281369] env[63175]: value = "task-1248340" [ 923.281369] env[63175]: _type = "Task" [ 923.281369] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.290846] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248340, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.321926] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.322209] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.322536] env[63175]: DEBUG nova.objects.instance [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.337763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.338111] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.338470] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.338751] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.339017] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.341439] env[63175]: INFO nova.compute.manager [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Terminating instance [ 923.362047] env[63175]: DEBUG nova.network.neutron [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.661763] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4286c2-63fa-4ecc-bcdb-e028b9b3ec1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.670406] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Suspending the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 923.670614] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-2855a2f1-b28e-45e8-b429-c7ea5aab99bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.677310] env[63175]: DEBUG oslo_vmware.api [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 923.677310] env[63175]: value = "task-1248341" [ 923.677310] env[63175]: _type = "Task" [ 923.677310] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.686769] env[63175]: DEBUG oslo_vmware.api [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248341, 'name': SuspendVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.792482] env[63175]: DEBUG oslo_vmware.api [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248340, 'name': RemoveSnapshot_Task, 'duration_secs': 0.402433} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.792807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 923.793090] env[63175]: INFO nova.compute.manager [None req-89397f63-f028-4766-99ce-2a730b343cde tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Took 12.30 seconds to snapshot the instance on the hypervisor. [ 923.827993] env[63175]: DEBUG nova.objects.instance [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.846042] env[63175]: DEBUG nova.compute.manager [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 923.846309] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.847277] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c64a58-b2a5-488c-a7b4-f65ede6e3061 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.856333] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.856638] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8a1b8cc-59bb-4633-b82d-454d91a74a32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.865567] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.867215] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 923.867215] env[63175]: value = "task-1248342" [ 923.867215] env[63175]: _type = "Task" [ 923.867215] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.881016] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.188262] env[63175]: DEBUG oslo_vmware.api [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248341, 'name': SuspendVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.331360] env[63175]: DEBUG nova.objects.base [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 924.331845] env[63175]: DEBUG nova.network.neutron [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.386145] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248342, 'name': PowerOffVM_Task, 'duration_secs': 0.192852} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.386367] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.386634] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.386980] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42b0be6d-ec5f-4b70-a3c4-fafd6ddf9c5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.469303] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.469730] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.470022] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleting the datastore file [datastore2] 9b72d607-be0f-4caa-b9e2-b25bbbe30f54 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.470253] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60ad6144-42ed-4358-8a9f-05c7abc32c5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.473639] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c8cee5c0-1230-4093-8a07-9629e8fafaf4 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.151s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.481103] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 924.481103] env[63175]: value = "task-1248344" [ 924.481103] env[63175]: _type = "Task" [ 924.481103] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.492520] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248344, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.631215] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.631496] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.631708] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.631906] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.632118] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.634379] env[63175]: INFO nova.compute.manager [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Terminating instance [ 924.688023] env[63175]: DEBUG oslo_vmware.api [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248341, 'name': SuspendVM_Task, 'duration_secs': 0.888911} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.688311] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Suspended the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 924.688495] env[63175]: DEBUG nova.compute.manager [None req-b0819f99-a4d0-4a55-81e0-d08e57e01759 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.689276] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c604af-fb0e-432b-8330-fcfa7caa3031 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.992421] env[63175]: DEBUG oslo_vmware.api [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248344, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242471} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.993251] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.993454] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.993733] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.993975] env[63175]: INFO nova.compute.manager [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Took 1.15 seconds to destroy the instance on the hypervisor. [ 924.994296] env[63175]: DEBUG oslo.service.loopingcall [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.994888] env[63175]: DEBUG nova.compute.manager [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 924.994993] env[63175]: DEBUG nova.network.neutron [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.138859] env[63175]: DEBUG nova.compute.manager [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 925.139123] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.140498] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07acedc1-6582-484f-b5d2-cf16b45d9a6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.148560] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.148827] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72dd72c5-553b-472b-ae7d-c4af72411683 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.211213] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.211475] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.235522] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.235794] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.236091] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore2] 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.236419] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bc09092-7a76-469e-846e-61330fc3bf9d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.243932] env[63175]: DEBUG oslo_vmware.api [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 925.243932] env[63175]: value = "task-1248346" [ 925.243932] env[63175]: _type = "Task" [ 925.243932] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.252756] env[63175]: DEBUG oslo_vmware.api [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.387505] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe22ca31-e2db-46ff-bdb9-19dc98ab9092 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.408110] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 0 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 925.715349] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 925.734478] env[63175]: DEBUG nova.compute.manager [req-25cc6aef-d4a0-4d3d-be86-bf1fad5e53ba req-6e5d2a47-4951-4a41-ac84-81db79d1ac45 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Received event network-vif-deleted-fcbca53b-aabe-41c3-bfdb-0821599b70d8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 925.734723] env[63175]: INFO nova.compute.manager [req-25cc6aef-d4a0-4d3d-be86-bf1fad5e53ba req-6e5d2a47-4951-4a41-ac84-81db79d1ac45 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Neutron deleted interface fcbca53b-aabe-41c3-bfdb-0821599b70d8; detaching it from the instance and deleting it from the info cache [ 925.734899] env[63175]: DEBUG nova.network.neutron [req-25cc6aef-d4a0-4d3d-be86-bf1fad5e53ba req-6e5d2a47-4951-4a41-ac84-81db79d1ac45 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.755148] env[63175]: DEBUG oslo_vmware.api [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170019} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.755379] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.755569] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.755751] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.755934] env[63175]: INFO nova.compute.manager [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Took 0.62 seconds to destroy the instance on the hypervisor. [ 925.756194] env[63175]: DEBUG oslo.service.loopingcall [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.756393] env[63175]: DEBUG nova.compute.manager [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 925.756486] env[63175]: DEBUG nova.network.neutron [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.916761] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.916761] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0af535e1-f830-4de2-b958-3c95f4a1f9e8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.925329] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 925.925329] env[63175]: value = "task-1248347" [ 925.925329] env[63175]: _type = "Task" [ 925.925329] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.940498] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.122391] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 926.122923] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 926.123155] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 926.212241] env[63175]: DEBUG nova.network.neutron [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.236229] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.236494] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.238161] env[63175]: INFO nova.compute.claims [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.240809] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4717678-f71e-4362-9852-b5a79431e3fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.250879] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf9249c-f644-43d9-8bdd-50c264761a90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.281302] env[63175]: DEBUG nova.compute.manager [req-25cc6aef-d4a0-4d3d-be86-bf1fad5e53ba req-6e5d2a47-4951-4a41-ac84-81db79d1ac45 service nova] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Detach interface failed, port_id=fcbca53b-aabe-41c3-bfdb-0821599b70d8, reason: Instance 9b72d607-be0f-4caa-b9e2-b25bbbe30f54 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 926.417400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.417776] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.418031] env[63175]: DEBUG nova.objects.instance [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.436343] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248347, 'name': PowerOffVM_Task, 'duration_secs': 0.199098} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.436532] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.436731] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 17 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 926.462288] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "ed618dba-1b54-4119-a317-4f64a64a59c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.462549] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.462768] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.462957] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.463147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.466306] env[63175]: INFO nova.compute.manager [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Terminating instance [ 926.531073] env[63175]: DEBUG nova.network.neutron [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.715108] env[63175]: INFO nova.compute.manager [-] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Took 1.72 seconds to deallocate network for instance. [ 926.775337] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.775581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.943063] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.943063] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.943338] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.943338] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.943468] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.943649] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.943869] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.944055] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.944236] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.944408] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.944604] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.949614] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ec3c82f-ef4e-4649-971d-30e5c2d235fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.970093] env[63175]: DEBUG nova.compute.manager [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 926.970322] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.970658] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 926.970658] env[63175]: value = "task-1248348" [ 926.970658] env[63175]: _type = "Task" [ 926.970658] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.971382] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c79772-b3a4-4a19-aefd-5ed44fb0a5c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.982724] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248348, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.984867] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.985138] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2897e852-de2d-4f55-8870-28ad4addd907 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.000818] env[63175]: DEBUG nova.objects.instance [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.033314] env[63175]: INFO nova.compute.manager [-] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Took 1.28 seconds to deallocate network for instance. [ 927.056474] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.056702] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.056900] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleting the datastore file [datastore1] ed618dba-1b54-4119-a317-4f64a64a59c9 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.057208] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cce7617a-657a-44f8-a059-8f4d561fb728 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.064719] env[63175]: DEBUG oslo_vmware.api [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 927.064719] env[63175]: value = "task-1248350" [ 927.064719] env[63175]: _type = "Task" [ 927.064719] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.076283] env[63175]: DEBUG oslo_vmware.api [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.164078] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.164266] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquired lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.164420] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Forcefully refreshing network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 927.221114] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.278555] env[63175]: DEBUG nova.compute.utils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 927.394491] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533bc11a-d3e7-405b-8b1f-0e778b9cbc57 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.401267] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29608d12-ca6e-4319-bac6-db2fc887be70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.432155] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb3fa04-6f78-4c72-bc63-09f3779137b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.440157] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172656b1-1983-43ea-9542-3854e9d38497 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.455235] env[63175]: DEBUG nova.compute.provider_tree [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.483483] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248348, 'name': ReconfigVM_Task, 'duration_secs': 0.179029} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.483843] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 33 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 927.504520] env[63175]: DEBUG nova.objects.base [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 927.504520] env[63175]: DEBUG nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 927.539907] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.541785] env[63175]: DEBUG nova.policy [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.576303] env[63175]: DEBUG oslo_vmware.api [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142037} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.576553] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.576795] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.576986] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.577300] env[63175]: INFO nova.compute.manager [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Took 0.61 seconds to destroy the instance on the hypervisor. [ 927.577502] env[63175]: DEBUG oslo.service.loopingcall [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.577709] env[63175]: DEBUG nova.compute.manager [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 927.577808] env[63175]: DEBUG nova.network.neutron [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.781045] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.933919] env[63175]: DEBUG nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Successfully created port: b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.960030] env[63175]: DEBUG nova.scheduler.client.report [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 927.990009] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.990557] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.991738] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.991738] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.991738] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.991738] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.991738] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.992170] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.992524] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.992821] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.993128] env[63175]: DEBUG nova.virt.hardware [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.998634] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 927.999302] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-019268e0-7373-4ef0-b9b9-82653e77d4ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.019879] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 928.019879] env[63175]: value = "task-1248351" [ 928.019879] env[63175]: _type = "Task" [ 928.019879] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.028997] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248351, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.376903] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [{"id": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "address": "fa:16:3e:cc:ea:d9", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71756e11-f6", "ovs_interfaceid": "71756e11-f67d-4268-a4b0-25b0a8cdefbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.412361] env[63175]: DEBUG nova.network.neutron [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.463420] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.227s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.464259] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 928.467029] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.246s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.467165] env[63175]: DEBUG nova.objects.instance [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'resources' on Instance uuid 9b72d607-be0f-4caa-b9e2-b25bbbe30f54 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.530663] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248351, 'name': ReconfigVM_Task, 'duration_secs': 0.162576} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.531186] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.532138] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4915998-a89e-4715-9dfa-2c5f52bd39ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.557585] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.557950] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fb0f6b4-0004-4045-8aa8-398458b5a1fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.578187] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 928.578187] env[63175]: value = "task-1248352" [ 928.578187] env[63175]: _type = "Task" [ 928.578187] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.587318] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248352, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.619816] env[63175]: DEBUG nova.compute.manager [req-d220e397-f176-4217-a480-652772707bac req-4fc41ac7-f214-435d-a3fa-b4abeb26dd2a service nova] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Received event network-vif-deleted-eaec69ba-7e1a-4abf-b799-33836e2e6c04 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 928.852864] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.853172] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.853478] env[63175]: INFO nova.compute.manager [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Attaching volume 345211ba-3866-4941-b49b-db275c459229 to /dev/sdb [ 928.879950] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Releasing lock "refresh_cache-cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.880151] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updated the network info_cache for instance {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 928.880418] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.880721] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.880979] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.881240] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.881501] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.881757] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.881941] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 928.882106] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.885250] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8663d8-89aa-4d03-9ec5-81f6aa27e5aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.894488] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6323e78a-4c6e-46bb-9d69-0880ee1931f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.909493] env[63175]: DEBUG nova.virt.block_device [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating existing volume attachment record: f208b46f-80c8-477d-9600-7cefb7cd40c9 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 928.915426] env[63175]: INFO nova.compute.manager [-] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Took 1.34 seconds to deallocate network for instance. [ 928.970300] env[63175]: DEBUG nova.compute.utils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.974517] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 928.974721] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.023125] env[63175]: DEBUG nova.policy [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e43774590162406a981cf0d5e20f5a1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83fa8f008a48484195c88ed76ceece0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.089248] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248352, 'name': ReconfigVM_Task, 'duration_secs': 0.266291} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.091847] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Reconfigured VM instance instance-00000051 to attach disk [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc/aad1653a-43a7-4637-acd5-7dbf76ea1ccc.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.092161] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 50 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.149629] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a03ac1-59a1-4e32-a1a0-d8706faf5f46 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.157272] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd34527c-a7a1-4165-ad2b-7c155fb78745 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.191127] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1308fb-3880-4885-996d-bdbc1a514a5c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.199085] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7edf1af-a601-42f5-a949-f53cea33e30d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.213392] env[63175]: DEBUG nova.compute.provider_tree [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.356773] env[63175]: DEBUG nova.compute.manager [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-plugged-b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 929.357103] env[63175]: DEBUG oslo_concurrency.lockutils [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.357779] env[63175]: DEBUG oslo_concurrency.lockutils [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.358246] env[63175]: DEBUG oslo_concurrency.lockutils [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.359166] env[63175]: DEBUG nova.compute.manager [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] No waiting events found dispatching network-vif-plugged-b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 929.359433] env[63175]: WARNING nova.compute.manager [req-9b2d99be-b071-4307-84c2-d2e1977a9d5f req-bb4f7567-8dfa-463b-88b6-0e6ae3a41b26 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received unexpected event network-vif-plugged-b75f0d56-f89a-4754-a893-e06356ede92f for instance with vm_state active and task_state None. [ 929.388534] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Successfully created port: 398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.393018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.422151] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.475840] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 929.599493] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effb1fb0-e9e1-4dfd-bcb9-7295bf602a87 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.625225] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f95cfbc-90bf-43e1-8ff7-7f1b1973698a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.646501] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 67 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.716793] env[63175]: DEBUG nova.scheduler.client.report [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.968623] env[63175]: DEBUG nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Successfully updated port: b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.185855] env[63175]: DEBUG nova.network.neutron [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Port 95854088-699b-4fec-a714-1b6959135fd2 binding to destination host cpu-1 is already ACTIVE {{(pid=63175) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 930.220929] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.223179] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.683s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.223416] env[63175]: DEBUG nova.objects.instance [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'resources' on Instance uuid 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.242030] env[63175]: INFO nova.scheduler.client.report [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted allocations for instance 9b72d607-be0f-4caa-b9e2-b25bbbe30f54 [ 930.471388] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.471511] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.471680] env[63175]: DEBUG nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.485837] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 930.509881] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.510360] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.510360] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.510509] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.510684] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.510865] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.511183] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.511437] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.511690] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.511880] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.512076] env[63175]: DEBUG nova.virt.hardware [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.512917] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0f8f9d-f85f-440b-9a86-8eda1cb15512 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.521737] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d27f93-2926-4d5c-9edd-ad720009718f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.730745] env[63175]: DEBUG nova.compute.manager [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Received event network-vif-deleted-4183e5b2-8de0-4874-a2b8-6a15994712b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 930.730975] env[63175]: DEBUG nova.compute.manager [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-changed-b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 930.731226] env[63175]: DEBUG nova.compute.manager [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing instance network info cache due to event network-changed-b75f0d56-f89a-4754-a893-e06356ede92f. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 930.731565] env[63175]: DEBUG oslo_concurrency.lockutils [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.749540] env[63175]: DEBUG oslo_concurrency.lockutils [None req-97dade4e-cfb7-4767-a0f7-bee52ed10757 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "9b72d607-be0f-4caa-b9e2-b25bbbe30f54" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.411s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.880273] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d938d5-263a-4f39-838e-082730fc6c1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.891179] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba96e3e3-f1d1-466b-a1ba-bc524a0c286d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.926489] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Successfully updated port: 398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.929606] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f19625e-ccbf-4f11-aaab-228c03b0c0df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.938670] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37ca844-c5c3-4cfd-8ee3-f7fba24e7acc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.953953] env[63175]: DEBUG nova.compute.provider_tree [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.008214] env[63175]: WARNING nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 931.210043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.210043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.210263] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.283283] env[63175]: DEBUG nova.network.neutron [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b75f0d56-f89a-4754-a893-e06356ede92f", "address": "fa:16:3e:9f:f2:d5", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb75f0d56-f8", "ovs_interfaceid": "b75f0d56-f89a-4754-a893-e06356ede92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.429514] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.429663] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.429819] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.456756] env[63175]: DEBUG nova.scheduler.client.report [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 931.786918] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.787666] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.787879] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.788185] env[63175]: DEBUG oslo_concurrency.lockutils [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.788380] env[63175]: DEBUG nova.network.neutron [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing network info cache for port b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.790300] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3145cc-a17f-4875-a666-3f821b1edabd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.807904] env[63175]: DEBUG oslo_concurrency.lockutils [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.808169] env[63175]: DEBUG oslo_concurrency.lockutils [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.809742] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.809979] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.810153] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.810351] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.810507] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.810660] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.810909] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.811106] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.811286] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.811459] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.811638] env[63175]: DEBUG nova.virt.hardware [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.819801] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfiguring VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 931.821048] env[63175]: INFO nova.compute.manager [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Detaching volume 226aa373-bf1a-46ed-b3fe-5461541d261b [ 931.822978] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e386635b-9847-4f3e-ae55-a7f207bab9aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.840852] env[63175]: DEBUG oslo_vmware.api [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 931.840852] env[63175]: value = "task-1248357" [ 931.840852] env[63175]: _type = "Task" [ 931.840852] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.849382] env[63175]: DEBUG oslo_vmware.api [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248357, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.864497] env[63175]: INFO nova.virt.block_device [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Attempting to driver detach volume 226aa373-bf1a-46ed-b3fe-5461541d261b from mountpoint /dev/sdb [ 931.864811] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 931.865021] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269045', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'name': 'volume-226aa373-bf1a-46ed-b3fe-5461541d261b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e4f169-000c-4e9c-8ef5-aa4b4989eb44', 'attached_at': '', 'detached_at': '', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'serial': '226aa373-bf1a-46ed-b3fe-5461541d261b'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 931.865852] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0caf8d-1470-45fa-87c8-8504fccaed23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.887355] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbba386-1f26-4bdc-8e82-28d54388d1d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.895430] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4349e8bc-af50-43bd-9a46-6ebfa508ff6e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.917034] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d8d691-e593-423c-9b13-8d3d61e2653c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.935116] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] The volume has not been displaced from its original location: [datastore2] volume-226aa373-bf1a-46ed-b3fe-5461541d261b/volume-226aa373-bf1a-46ed-b3fe-5461541d261b.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 931.940307] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 931.940673] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cbdb0fa-22f2-44f0-91f9-c13f7f63daad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.961991] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.964087] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 931.964087] env[63175]: value = "task-1248358" [ 931.964087] env[63175]: _type = "Task" [ 931.964087] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.965403] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.573s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.965403] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.965403] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 931.965403] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.543s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.965403] env[63175]: DEBUG nova.objects.instance [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'resources' on Instance uuid ed618dba-1b54-4119-a317-4f64a64a59c9 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.969046] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117b8e24-1759-4062-a790-4d15969cfe3e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.980983] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248358, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.983212] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b20ed6-2706-40c7-bca5-50cc73a0fc93 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.988346] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.992251] env[63175]: INFO nova.scheduler.client.report [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted allocations for instance 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8 [ 932.012022] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eab9b4e-126e-4433-88b9-82f38571b3c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.018408] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359406e6-512c-402c-bc95-571af264b2ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.051924] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180118MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 932.052108] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.169311] env[63175]: DEBUG nova.network.neutron [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Updating instance_info_cache with network_info: [{"id": "398121e4-74d6-495f-a429-da631d9be9a5", "address": "fa:16:3e:b2:a3:15", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap398121e4-74", "ovs_interfaceid": "398121e4-74d6-495f-a429-da631d9be9a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.241525] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.241738] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.241966] env[63175]: DEBUG nova.network.neutron [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.351451] env[63175]: DEBUG oslo_vmware.api [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248357, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.479268] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248358, 'name': ReconfigVM_Task, 'duration_secs': 0.236279} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.479563] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 932.484385] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39510d14-d872-452c-a061-3782b4c67e2f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.504286] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 932.504286] env[63175]: value = "task-1248359" [ 932.504286] env[63175]: _type = "Task" [ 932.504286] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.508965] env[63175]: DEBUG nova.network.neutron [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updated VIF entry in instance network info cache for port b75f0d56-f89a-4754-a893-e06356ede92f. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.509502] env[63175]: DEBUG nova.network.neutron [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b75f0d56-f89a-4754-a893-e06356ede92f", "address": "fa:16:3e:9f:f2:d5", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb75f0d56-f8", "ovs_interfaceid": "b75f0d56-f89a-4754-a893-e06356ede92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.516044] env[63175]: DEBUG oslo_concurrency.lockutils [None req-5e0854a9-5b8f-4ca6-8de5-5487608086d7 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.884s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.525523] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.628487] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eed814a-fa7c-4f74-aa83-40623f9b64fe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.637918] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba2418b-7195-4932-9af9-649bf2e6d4f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.670850] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a08cef-0fab-49b9-8aa9-0e3b2a6bc1f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.673715] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.674024] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Instance network_info: |[{"id": "398121e4-74d6-495f-a429-da631d9be9a5", "address": "fa:16:3e:b2:a3:15", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap398121e4-74", "ovs_interfaceid": "398121e4-74d6-495f-a429-da631d9be9a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 932.674411] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:a3:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '721c6720-3ce0-450e-9951-a894f03acc27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '398121e4-74d6-495f-a429-da631d9be9a5', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.681890] env[63175]: DEBUG oslo.service.loopingcall [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.682503] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 932.682744] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ac5a9e2-bb83-4b3c-b16a-9974606b20b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.699935] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dfcbd4-dc1c-4fe3-9d83-3e39bb486a0f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.704605] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.704605] env[63175]: value = "task-1248360" [ 932.704605] env[63175]: _type = "Task" [ 932.704605] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.727508] env[63175]: DEBUG nova.compute.provider_tree [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.731376] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248360, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.758166] env[63175]: DEBUG nova.compute.manager [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Received event network-vif-plugged-398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 932.758405] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Acquiring lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.758625] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.758826] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.759017] env[63175]: DEBUG nova.compute.manager [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] No waiting events found dispatching network-vif-plugged-398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 932.759263] env[63175]: WARNING nova.compute.manager [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Received unexpected event network-vif-plugged-398121e4-74d6-495f-a429-da631d9be9a5 for instance with vm_state building and task_state spawning. [ 932.759367] env[63175]: DEBUG nova.compute.manager [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Received event network-changed-398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 932.759562] env[63175]: DEBUG nova.compute.manager [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Refreshing instance network info cache due to event network-changed-398121e4-74d6-495f-a429-da631d9be9a5. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 932.759853] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Acquiring lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.760036] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Acquired lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.760212] env[63175]: DEBUG nova.network.neutron [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Refreshing network info cache for port 398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.851489] env[63175]: DEBUG oslo_vmware.api [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248357, 'name': ReconfigVM_Task, 'duration_secs': 0.620652} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.854105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.854347] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfigured VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 932.968374] env[63175]: DEBUG nova.network.neutron [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.017783] env[63175]: DEBUG oslo_concurrency.lockutils [req-ea529ed9-6776-4907-b243-f11a3684b4f0 req-faa95d5c-2452-4012-899b-d85ff5d293b4 service nova] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.018372] env[63175]: DEBUG oslo_vmware.api [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248359, 'name': ReconfigVM_Task, 'duration_secs': 0.153427} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.019029] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269045', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'name': 'volume-226aa373-bf1a-46ed-b3fe-5461541d261b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e4f169-000c-4e9c-8ef5-aa4b4989eb44', 'attached_at': '', 'detached_at': '', 'volume_id': '226aa373-bf1a-46ed-b3fe-5461541d261b', 'serial': '226aa373-bf1a-46ed-b3fe-5461541d261b'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 933.215992] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248360, 'name': CreateVM_Task, 'duration_secs': 0.326535} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.216225] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 933.216877] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.217064] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.217392] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.217651] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111aabfe-ebf3-4d75-be3d-6def200051c4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.222168] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 933.222168] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529532aa-7b40-a6e5-45ce-ded60670f715" [ 933.222168] env[63175]: _type = "Task" [ 933.222168] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.229999] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529532aa-7b40-a6e5-45ce-ded60670f715, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.233256] env[63175]: DEBUG nova.scheduler.client.report [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 933.358702] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6e6f5a25-f13e-485b-938a-99e542b41b5d tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.941s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.453105] env[63175]: DEBUG nova.network.neutron [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Updated VIF entry in instance network info cache for port 398121e4-74d6-495f-a429-da631d9be9a5. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.453210] env[63175]: DEBUG nova.network.neutron [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Updating instance_info_cache with network_info: [{"id": "398121e4-74d6-495f-a429-da631d9be9a5", "address": "fa:16:3e:b2:a3:15", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap398121e4-74", "ovs_interfaceid": "398121e4-74d6-495f-a429-da631d9be9a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.456687] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 933.456910] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269071', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'name': 'volume-345211ba-3866-4941-b49b-db275c459229', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0645fffe-9910-43af-af40-126592aefaff', 'attached_at': '', 'detached_at': '', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'serial': '345211ba-3866-4941-b49b-db275c459229'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 933.457750] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77a1e59-9a26-485a-997e-94f26ccd8204 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.474552] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.478653] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d800de38-5146-4571-bac1-f707ae12e1b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.504773] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-345211ba-3866-4941-b49b-db275c459229/volume-345211ba-3866-4941-b49b-db275c459229.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.505198] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-887f9f73-8c9a-4cd8-9798-4d350c75fe5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.527553] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 933.527553] env[63175]: value = "task-1248361" [ 933.527553] env[63175]: _type = "Task" [ 933.527553] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.536577] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.563776] env[63175]: DEBUG nova.objects.instance [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.732519] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529532aa-7b40-a6e5-45ce-ded60670f715, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.732770] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.733028] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.733280] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.733436] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.733624] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.733900] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7a9192a-8ede-45d9-bda1-7af0257a80a0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.737575] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.772s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.739871] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.688s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.743047] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.743047] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.743442] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8e9e728-2a43-456b-baba-28dbf1c9e79f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.748941] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 933.748941] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521db4fa-1d23-80cd-541b-7c5e4c470717" [ 933.748941] env[63175]: _type = "Task" [ 933.748941] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.759480] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521db4fa-1d23-80cd-541b-7c5e4c470717, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.760433] env[63175]: INFO nova.scheduler.client.report [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance ed618dba-1b54-4119-a317-4f64a64a59c9 [ 933.956554] env[63175]: DEBUG oslo_concurrency.lockutils [req-3d5cce38-2f95-4ad4-9a4e-19cecc0e9be4 req-95968959-d77e-4476-ad9a-292efba695b4 service nova] Releasing lock "refresh_cache-9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.999110] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a0f387-a511-49e5-ac30-1fa0b57570c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.020264] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95772dbc-f8bd-4cb2-93cd-ffd4399a793f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.028367] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 83 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 934.042249] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248361, 'name': ReconfigVM_Task, 'duration_secs': 0.347086} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.042584] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-345211ba-3866-4941-b49b-db275c459229/volume-345211ba-3866-4941-b49b-db275c459229.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.047403] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f00a9ac-594d-4e0e-b5f9-3c856faeb5b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.063617] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 934.063617] env[63175]: value = "task-1248362" [ 934.063617] env[63175]: _type = "Task" [ 934.063617] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.075774] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248362, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.260861] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]521db4fa-1d23-80cd-541b-7c5e4c470717, 'name': SearchDatastore_Task, 'duration_secs': 0.010422} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.261949] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ff8084b-efd1-4274-a3e2-d4b78385448c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.267870] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efc4d938-e4d4-4a5a-86a8-e3132aede6eb tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "ed618dba-1b54-4119-a317-4f64a64a59c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.805s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.270889] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 934.270889] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525cadde-a068-93ae-c81b-d1e9bcfc70b9" [ 934.270889] env[63175]: _type = "Task" [ 934.270889] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.279647] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525cadde-a068-93ae-c81b-d1e9bcfc70b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.539287] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.539587] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2092ccd3-286a-42da-8da3-37a7e58ed5bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.548599] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 934.548599] env[63175]: value = "task-1248363" [ 934.548599] env[63175]: _type = "Task" [ 934.548599] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.556874] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248363, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.573105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-291874f8-2639-4570-a923-50df80a65516 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.765s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.574156] env[63175]: DEBUG oslo_vmware.api [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248362, 'name': ReconfigVM_Task, 'duration_secs': 0.161775} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.574493] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269071', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'name': 'volume-345211ba-3866-4941-b49b-db275c459229', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0645fffe-9910-43af-af40-126592aefaff', 'attached_at': '', 'detached_at': '', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'serial': '345211ba-3866-4941-b49b-db275c459229'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 934.756047] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Applying migration context for instance aad1653a-43a7-4637-acd5-7dbf76ea1ccc as it has an incoming, in-progress migration 757bbb05-bc15-40f2-b008-b8179561f844. Migration status is post-migrating {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 934.757679] env[63175]: INFO nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating resource usage from migration 757bbb05-bc15-40f2-b008-b8179561f844 [ 934.775415] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.775581] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.775681] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a460926e-9637-40aa-bb30-e3890a441e03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.775799] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.775915] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a4304579-f829-433e-a878-1050fac08c2e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.776155] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0645fffe-9910-43af-af40-126592aefaff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.776312] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.776431] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Migration 757bbb05-bc15-40f2-b008-b8179561f844 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 934.776547] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance aad1653a-43a7-4637-acd5-7dbf76ea1ccc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.776657] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 934.784188] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525cadde-a068-93ae-c81b-d1e9bcfc70b9, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.784757] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.785065] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db/9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.785337] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55b4abb9-d7ea-4e71-bd9f-09043800c35f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.793319] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 934.793319] env[63175]: value = "task-1248364" [ 934.793319] env[63175]: _type = "Task" [ 934.793319] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.801731] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.948123] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.948443] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.005080] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-a4304579-f829-433e-a878-1050fac08c2e-c7e98f4d-1b20-4610-ae0d-74b0c13618ba" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.005390] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-c7e98f4d-1b20-4610-ae0d-74b0c13618ba" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.005513] env[63175]: DEBUG nova.objects.instance [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.025194] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.025493] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.025819] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.026024] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.026086] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.028309] env[63175]: INFO nova.compute.manager [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Terminating instance [ 935.060189] env[63175]: DEBUG oslo_vmware.api [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248363, 'name': PowerOnVM_Task, 'duration_secs': 0.454847} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.060491] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.060689] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8110dc06-2a0c-43a8-9c3d-e36c523bcc16 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance 'aad1653a-43a7-4637-acd5-7dbf76ea1ccc' progress to 100 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 935.279961] env[63175]: INFO nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0633c8e9-1c18-4047-ae5d-012d715e03df has allocations against this compute host but is not found in the database. [ 935.280280] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 935.280419] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 935.303575] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248364, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.420909] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b812c5-4ce2-4aa8-931f-5e9110ea3861 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.428926] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0e7a21-bbe6-4196-a88b-065d76801e23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.460291] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 935.463924] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1c04f5-6618-4825-be07-2b1db3c5c3c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.471614] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b7c299-30b4-475e-a636-66a92aaa08bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.485464] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.532637] env[63175]: DEBUG nova.compute.manager [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 935.532868] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.535867] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d61549-b521-468d-b3ff-6ef27737d666 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.544321] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.544610] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a255c12d-d8fd-4113-9c16-be38e974b0b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.551805] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 935.551805] env[63175]: value = "task-1248365" [ 935.551805] env[63175]: _type = "Task" [ 935.551805] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.559799] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248365, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.607828] env[63175]: DEBUG nova.objects.instance [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.619727] env[63175]: DEBUG nova.objects.instance [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'flavor' on Instance uuid 0645fffe-9910-43af-af40-126592aefaff {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.804182] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248364, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515879} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.804479] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db/9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.804704] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.804968] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da1a427e-2101-47f1-908a-381599b8e1a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.811957] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 935.811957] env[63175]: value = "task-1248366" [ 935.811957] env[63175]: _type = "Task" [ 935.811957] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.820512] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248366, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.977874] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.988192] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 936.061814] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248365, 'name': PowerOffVM_Task, 'duration_secs': 0.243956} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.062157] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.062342] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 936.062592] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a76cf4f-0285-48e3-bc56-28fee0a11346 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.110543] env[63175]: DEBUG nova.objects.base [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 936.110773] env[63175]: DEBUG nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.124995] env[63175]: DEBUG oslo_concurrency.lockutils [None req-0fdfa518-f499-49ea-af2a-586f01c3e0aa tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.272s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.128038] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.128328] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.128530] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleting the datastore file [datastore2] e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.129101] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e334519c-a5b9-4814-8679-3cd7c2cb820c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.136537] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 936.136537] env[63175]: value = "task-1248368" [ 936.136537] env[63175]: _type = "Task" [ 936.136537] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.147394] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.176396] env[63175]: DEBUG nova.policy [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.322653] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248366, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062073} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.322957] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.323828] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e5e2ca-5975-4da2-8f16-c79e025f21a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.346176] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db/9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.346493] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc891c5a-2247-40e7-a1a0-db791773f5ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.360451] env[63175]: INFO nova.compute.manager [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Rescuing [ 936.360692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.360852] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.361037] env[63175]: DEBUG nova.network.neutron [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.368157] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 936.368157] env[63175]: value = "task-1248369" [ 936.368157] env[63175]: _type = "Task" [ 936.368157] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.377273] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248369, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.493118] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 936.493336] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.754s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.493622] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.516s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.495238] env[63175]: INFO nova.compute.claims [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.646579] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.878221] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248369, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.100178] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.100178] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.100178] env[63175]: DEBUG nova.compute.manager [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Going to confirm migration 2 {{(pid=63175) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 937.142316] env[63175]: DEBUG nova.network.neutron [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.149163] env[63175]: DEBUG oslo_vmware.api [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.832293} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.149338] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.149530] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.149724] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.149908] env[63175]: INFO nova.compute.manager [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Took 1.62 seconds to destroy the instance on the hypervisor. [ 937.150204] env[63175]: DEBUG oslo.service.loopingcall [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.150408] env[63175]: DEBUG nova.compute.manager [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 937.150520] env[63175]: DEBUG nova.network.neutron [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 937.379197] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248369, 'name': ReconfigVM_Task, 'duration_secs': 0.7529} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.379472] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db/9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.380109] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c01d905b-08be-49bc-8127-a788210926da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.387536] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 937.387536] env[63175]: value = "task-1248370" [ 937.387536] env[63175]: _type = "Task" [ 937.387536] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.395677] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248370, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.647922] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.696134] env[63175]: DEBUG nova.compute.manager [req-78c37683-c31f-4207-a95a-7f5cff9f6ece req-1a24fd6c-9c75-4e94-b1f7-56bd68998031 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Received event network-vif-deleted-57912c9e-afb5-4a48-af4e-c569456da08b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 937.696351] env[63175]: INFO nova.compute.manager [req-78c37683-c31f-4207-a95a-7f5cff9f6ece req-1a24fd6c-9c75-4e94-b1f7-56bd68998031 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Neutron deleted interface 57912c9e-afb5-4a48-af4e-c569456da08b; detaching it from the instance and deleting it from the info cache [ 937.696521] env[63175]: DEBUG nova.network.neutron [req-78c37683-c31f-4207-a95a-7f5cff9f6ece req-1a24fd6c-9c75-4e94-b1f7-56bd68998031 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.700725] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3701fa-2f92-414c-9051-9e48adcc688b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.705753] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.705942] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.706174] env[63175]: DEBUG nova.network.neutron [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.706380] env[63175]: DEBUG nova.objects.instance [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'info_cache' on Instance uuid aad1653a-43a7-4637-acd5-7dbf76ea1ccc {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.722728] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310c1876-d76f-4b71-902f-ffa9d88a966c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.756799] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb86331b-e29f-4dd0-8adb-e32f0f7cfc52 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.765157] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ef9bc8-d8ec-4dce-ad79-855fb383bbcb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.783569] env[63175]: DEBUG nova.compute.provider_tree [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.876098] env[63175]: DEBUG nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Successfully updated port: c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.900036] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248370, 'name': Rename_Task, 'duration_secs': 0.152489} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.900036] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.900036] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70d0fc66-9e46-4de7-87d2-4f788118695f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.908336] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 937.908336] env[63175]: value = "task-1248371" [ 937.908336] env[63175]: _type = "Task" [ 937.908336] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.917444] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248371, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.176084] env[63175]: DEBUG nova.network.neutron [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.205180] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75c555f8-5abe-4a50-9073-594ea5b74c51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.216982] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462eef11-f4ce-4d49-b879-ac4120d09481 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.246282] env[63175]: DEBUG nova.compute.manager [req-78c37683-c31f-4207-a95a-7f5cff9f6ece req-1a24fd6c-9c75-4e94-b1f7-56bd68998031 service nova] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Detach interface failed, port_id=57912c9e-afb5-4a48-af4e-c569456da08b, reason: Instance e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 938.287100] env[63175]: DEBUG nova.scheduler.client.report [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 938.378371] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.378584] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.379469] env[63175]: DEBUG nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.419500] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248371, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.678948] env[63175]: INFO nova.compute.manager [-] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Took 1.53 seconds to deallocate network for instance. [ 938.793301] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.793854] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 938.919894] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248371, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.933708] env[63175]: WARNING nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 938.934024] env[63175]: WARNING nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 939.159953] env[63175]: DEBUG nova.network.neutron [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [{"id": "95854088-699b-4fec-a714-1b6959135fd2", "address": "fa:16:3e:1e:d9:a7", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95854088-69", "ovs_interfaceid": "95854088-699b-4fec-a714-1b6959135fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.187331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.187331] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.187331] env[63175]: DEBUG nova.objects.instance [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'resources' on Instance uuid e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.195318] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.197303] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1facd623-ee41-4e24-bd0c-ae6ef6bbfb52 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.206521] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 939.206521] env[63175]: value = "task-1248372" [ 939.206521] env[63175]: _type = "Task" [ 939.206521] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.216282] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.299622] env[63175]: DEBUG nova.compute.utils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.301079] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 939.301268] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.367562] env[63175]: DEBUG nova.network.neutron [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b75f0d56-f89a-4754-a893-e06356ede92f", "address": "fa:16:3e:9f:f2:d5", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb75f0d56-f8", "ovs_interfaceid": "b75f0d56-f89a-4754-a893-e06356ede92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "address": "fa:16:3e:37:9d:cd", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7e98f4d-1b", "ovs_interfaceid": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.374456] env[63175]: DEBUG nova.policy [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89e51e3326e84b8c81358d205964bd1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71ca546385844c0d803034ef9e853377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.421452] env[63175]: DEBUG oslo_vmware.api [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248371, 'name': PowerOnVM_Task, 'duration_secs': 1.085307} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.421452] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.421650] env[63175]: INFO nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Took 8.94 seconds to spawn the instance on the hypervisor. [ 939.421838] env[63175]: DEBUG nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 939.423041] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1796ccdd-13f3-4dc8-9285-319f23e8e93d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.664182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-aad1653a-43a7-4637-acd5-7dbf76ea1ccc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.664182] env[63175]: DEBUG nova.objects.instance [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'migration_context' on Instance uuid aad1653a-43a7-4637-acd5-7dbf76ea1ccc {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.683770] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Successfully created port: 4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.723681] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248372, 'name': PowerOffVM_Task, 'duration_secs': 0.20264} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.726347] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.727435] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcec00f8-82a3-46c0-afbe-a1fafcf59550 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.750854] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2ae21e-483c-4edd-b9e0-7adc75acfb7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.795630] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.795937] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67b4c97f-58e3-4e8f-b766-371ff5a39b75 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.803976] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 939.807320] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 939.807320] env[63175]: value = "task-1248373" [ 939.807320] env[63175]: _type = "Task" [ 939.807320] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.823534] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 939.823810] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.824027] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.824191] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.824374] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.824816] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efd3717d-edae-4de6-8938-e8c6d1b73e68 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.833922] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.834137] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.837092] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f92d91f9-955d-4458-875d-1ec8351f61f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.844113] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 939.844113] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5201c3c5-4c18-f6ff-3793-aecfc66c903a" [ 939.844113] env[63175]: _type = "Task" [ 939.844113] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.852875] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5201c3c5-4c18-f6ff-3793-aecfc66c903a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.869845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.870469] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.870645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.871395] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36cfbbb-6924-4ab6-bbe9-2d9cd31bf038 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.889879] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.890202] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.890321] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.890513] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.890786] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.890882] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.891045] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.891293] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.891401] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.891580] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.892763] env[63175]: DEBUG nova.virt.hardware [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.899688] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfiguring VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 939.900939] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1da8cf49-6d07-4b1a-a9fd-af7380b8ff3b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.913316] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed956c16-9557-4d8f-a355-8d56f8a9c449 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.921905] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f69949-c87c-4897-aace-f4667d1ee48b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.926147] env[63175]: DEBUG oslo_vmware.api [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 939.926147] env[63175]: value = "task-1248374" [ 939.926147] env[63175]: _type = "Task" [ 939.926147] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.959244] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676814b7-a34d-4ea1-9fb1-3b4e76bb38e3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.967207] env[63175]: DEBUG oslo_vmware.api [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248374, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.967710] env[63175]: INFO nova.compute.manager [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Took 13.75 seconds to build instance. [ 939.975160] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825cd3fd-5143-48a1-9697-be449d266171 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.994691] env[63175]: DEBUG nova.compute.provider_tree [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.166283] env[63175]: DEBUG nova.objects.base [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 940.167288] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c8b55b-acf9-4501-bd55-c230a7b5a6a5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.189490] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc2c53b5-6bd2-4c97-8b32-f335d77ed092 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.196581] env[63175]: DEBUG oslo_vmware.api [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 940.196581] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526c6620-b67a-f243-d93e-a692ae938a17" [ 940.196581] env[63175]: _type = "Task" [ 940.196581] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.205639] env[63175]: DEBUG oslo_vmware.api [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526c6620-b67a-f243-d93e-a692ae938a17, 'name': SearchDatastore_Task} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.205923] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.284844] env[63175]: DEBUG nova.compute.manager [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-plugged-c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 940.284844] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.284844] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.284844] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.284987] env[63175]: DEBUG nova.compute.manager [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] No waiting events found dispatching network-vif-plugged-c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.285142] env[63175]: WARNING nova.compute.manager [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received unexpected event network-vif-plugged-c7e98f4d-1b20-4610-ae0d-74b0c13618ba for instance with vm_state active and task_state None. [ 940.285308] env[63175]: DEBUG nova.compute.manager [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-changed-c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 940.285467] env[63175]: DEBUG nova.compute.manager [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing instance network info cache due to event network-changed-c7e98f4d-1b20-4610-ae0d-74b0c13618ba. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 940.285750] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.285980] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.286146] env[63175]: DEBUG nova.network.neutron [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Refreshing network info cache for port c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.356069] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5201c3c5-4c18-f6ff-3793-aecfc66c903a, 'name': SearchDatastore_Task, 'duration_secs': 0.009696} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.356701] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81093c99-e5b4-4ff6-ad1d-c817dfc3616c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.362983] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 940.362983] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523af07b-7b0c-d23d-a8fe-3eb66a4d044d" [ 940.362983] env[63175]: _type = "Task" [ 940.362983] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.371067] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523af07b-7b0c-d23d-a8fe-3eb66a4d044d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.437544] env[63175]: DEBUG oslo_vmware.api [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.470491] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c7f0f263-73a9-48cb-b463-5bf890533e75 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.259s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.497414] env[63175]: DEBUG nova.scheduler.client.report [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 940.815010] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 940.836546] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.836807] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.836971] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.837186] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.837341] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.837495] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.837708] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.837871] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.838059] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.838235] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.838411] env[63175]: DEBUG nova.virt.hardware [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.839290] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7640517c-6c8a-4521-9f97-8441dbeafe9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.849224] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd3ef76-03d8-4226-a765-039d3b9fcfc6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.872869] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523af07b-7b0c-d23d-a8fe-3eb66a4d044d, 'name': SearchDatastore_Task, 'duration_secs': 0.009563} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.873549] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.873549] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0645fffe-9910-43af-af40-126592aefaff/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. {{(pid=63175) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 940.873742] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bef70487-7ee8-44ab-b65e-00debc0c63a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.880336] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 940.880336] env[63175]: value = "task-1248375" [ 940.880336] env[63175]: _type = "Task" [ 940.880336] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.890298] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248375, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.939048] env[63175]: DEBUG oslo_vmware.api [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248374, 'name': ReconfigVM_Task, 'duration_secs': 0.68331} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.942033] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.942214] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfigured VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 941.003127] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.005692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.800s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.015877] env[63175]: DEBUG nova.network.neutron [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updated VIF entry in instance network info cache for port c7e98f4d-1b20-4610-ae0d-74b0c13618ba. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.016708] env[63175]: DEBUG nova.network.neutron [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b75f0d56-f89a-4754-a893-e06356ede92f", "address": "fa:16:3e:9f:f2:d5", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb75f0d56-f8", "ovs_interfaceid": "b75f0d56-f89a-4754-a893-e06356ede92f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "address": "fa:16:3e:37:9d:cd", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7e98f4d-1b", "ovs_interfaceid": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.028739] env[63175]: INFO nova.scheduler.client.report [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted allocations for instance e1e4f169-000c-4e9c-8ef5-aa4b4989eb44 [ 941.268943] env[63175]: DEBUG nova.compute.manager [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.270712] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf5d9f7-e953-4cae-9eed-6238a43af19c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.305735] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Successfully updated port: 4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.392457] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248375, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455232} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.392683] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 0645fffe-9910-43af-af40-126592aefaff/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. [ 941.393449] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef8f8f3-819d-4eaa-92f7-cd9d7c9d2fff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.420665] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 0645fffe-9910-43af-af40-126592aefaff/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.421287] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70b5cae7-78de-4718-b71e-1adb0a984a87 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.441708] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 941.441708] env[63175]: value = "task-1248376" [ 941.441708] env[63175]: _type = "Task" [ 941.441708] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.447341] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f7ad6dde-97b1-43c5-94b5-fe2ec3e887d0 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-c7e98f4d-1b20-4610-ae0d-74b0c13618ba" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.442s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.451487] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.519348] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f39fba4-de42-478e-a7b3-da3fc6f779ae req-cdaa4c27-e75f-4191-a24d-fd288ceefeaa service nova] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.538105] env[63175]: DEBUG oslo_concurrency.lockutils [None req-67c0f113-9fee-4dfe-8877-5b6f8b366e6f tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "e1e4f169-000c-4e9c-8ef5-aa4b4989eb44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.512s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.648780] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58fa922-b396-49fd-8c6e-5f339a215cdc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.657763] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0deb11ed-3019-488d-ab1a-64462ea3fccc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.690414] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011104c1-cccf-44d3-b5cc-b97a28c6c7d3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.699199] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8958ce-917c-4eeb-8b0b-be950e598f92 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.714176] env[63175]: DEBUG nova.compute.provider_tree [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.784264] env[63175]: INFO nova.compute.manager [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] instance snapshotting [ 941.787394] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bef6db-72e6-44c5-a787-ed3d93e24370 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.811332] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.811506] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.811729] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.817252] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ca08e7-11b8-4ddb-8657-5d6c8574bc9f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.953866] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248376, 'name': ReconfigVM_Task, 'duration_secs': 0.358294} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.954183] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 0645fffe-9910-43af-af40-126592aefaff/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.955131] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a727dd54-f087-4759-ade4-fe01a89a0266 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.984567] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bb53840-fdbe-43bc-8032-727fd1fcaf7e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.003018] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 942.003018] env[63175]: value = "task-1248377" [ 942.003018] env[63175]: _type = "Task" [ 942.003018] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.011811] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.217830] env[63175]: DEBUG nova.scheduler.client.report [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 942.313299] env[63175]: DEBUG nova.compute.manager [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Received event network-vif-plugged-4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 942.313605] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.313770] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.314018] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.314656] env[63175]: DEBUG nova.compute.manager [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] No waiting events found dispatching network-vif-plugged-4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 942.314789] env[63175]: WARNING nova.compute.manager [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Received unexpected event network-vif-plugged-4621f750-c3ff-4289-bf4f-9ad952a25f05 for instance with vm_state building and task_state spawning. [ 942.314982] env[63175]: DEBUG nova.compute.manager [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Received event network-changed-4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 942.315215] env[63175]: DEBUG nova.compute.manager [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Refreshing instance network info cache due to event network-changed-4621f750-c3ff-4289-bf4f-9ad952a25f05. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 942.315411] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Acquiring lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.324832] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 942.325154] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-81e58da3-37ff-4ffa-8d02-4c646594e103 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.335032] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 942.335032] env[63175]: value = "task-1248378" [ 942.335032] env[63175]: _type = "Task" [ 942.335032] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.345864] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.348326] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.513743] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248377, 'name': ReconfigVM_Task, 'duration_secs': 0.169921} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.514081] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.514885] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5322d403-53b6-481c-abcf-0fa0a08d1cca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.522644] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 942.522644] env[63175]: value = "task-1248379" [ 942.522644] env[63175]: _type = "Task" [ 942.522644] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.532509] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.543497] env[63175]: DEBUG nova.network.neutron [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.843745] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248378, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.943180] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-a4304579-f829-433e-a878-1050fac08c2e-b75f0d56-f89a-4754-a893-e06356ede92f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.943483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-b75f0d56-f89a-4754-a893-e06356ede92f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.033773] env[63175]: DEBUG oslo_vmware.api [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248379, 'name': PowerOnVM_Task, 'duration_secs': 0.476668} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.034166] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.037047] env[63175]: DEBUG nova.compute.manager [None req-2f355b28-b4c7-45e8-9083-2d33f5453a1c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.037837] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a34d90-2ec7-4b83-8bfa-661328a647b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.046858] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.047152] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Instance network_info: |[{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 943.047600] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Acquired lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.047782] env[63175]: DEBUG nova.network.neutron [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Refreshing network info cache for port 4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.049488] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:43:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4621f750-c3ff-4289-bf4f-9ad952a25f05', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.056843] env[63175]: DEBUG oslo.service.loopingcall [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.061284] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.061891] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b8262da-3fad-4c71-a4d5-d77a5b4c52d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.084356] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.084356] env[63175]: value = "task-1248380" [ 943.084356] env[63175]: _type = "Task" [ 943.084356] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.093919] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248380, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.230704] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.225s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.313651] env[63175]: DEBUG nova.network.neutron [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updated VIF entry in instance network info cache for port 4621f750-c3ff-4289-bf4f-9ad952a25f05. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.314249] env[63175]: DEBUG nova.network.neutron [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.344146] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248378, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.447312] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.447514] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.448405] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4043547c-6036-48f6-bdc0-1144f9a4bec6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.470412] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a993fb-6d88-4783-b558-2f1fe5e0f94a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.499162] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfiguring VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 943.499495] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3daed62c-c562-44a3-918f-000f5581d851 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.519106] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 943.519106] env[63175]: value = "task-1248381" [ 943.519106] env[63175]: _type = "Task" [ 943.519106] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.528713] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.594498] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248380, 'name': CreateVM_Task, 'duration_secs': 0.501937} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.594677] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.595408] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.595583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.595908] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.596173] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35c0c6ec-b7f4-4405-84b5-5f60d8d65356 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.601176] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 943.601176] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527aedab-5014-49e3-ac63-60f872d5f3a6" [ 943.601176] env[63175]: _type = "Task" [ 943.601176] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.608898] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527aedab-5014-49e3-ac63-60f872d5f3a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.730107] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.730352] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.780592] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.780869] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.781100] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.781292] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.781776] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.784910] env[63175]: INFO nova.scheduler.client.report [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocation for migration 757bbb05-bc15-40f2-b008-b8179561f844 [ 943.785976] env[63175]: INFO nova.compute.manager [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Terminating instance [ 943.816572] env[63175]: DEBUG oslo_concurrency.lockutils [req-6a9ab323-395b-4344-9069-491abc347a0f req-296fd1c2-3f19-4f2a-bd39-c212cbcb631e service nova] Releasing lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.846627] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248378, 'name': CreateSnapshot_Task, 'duration_secs': 1.064848} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.846917] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 943.847709] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c743e09a-7f29-44d3-81bc-2d7bea9788bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.030784] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.118824] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527aedab-5014-49e3-ac63-60f872d5f3a6, 'name': SearchDatastore_Task, 'duration_secs': 0.036341} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.119436] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.119693] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.120037] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.120277] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.120702] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.121310] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d433990-9257-4160-8132-01f3f0214bd1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.131720] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.132110] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.132737] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a850b781-2828-41ea-93c0-53fe9d66f596 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.140204] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 944.140204] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f428c6-f8cd-7e95-a821-afd30d75702b" [ 944.140204] env[63175]: _type = "Task" [ 944.140204] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.147996] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f428c6-f8cd-7e95-a821-afd30d75702b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.233086] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 944.292641] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c502d3d8-b355-4ad7-b448-9731770e2397 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.193s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.294405] env[63175]: DEBUG nova.compute.manager [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 944.294405] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.295471] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6ad3bd-0ecf-47e0-b82f-e2c78c5cb9ae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.303429] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.303715] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23d36fc6-acc1-457f-b39e-72cd53b13cf8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.312215] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 944.312215] env[63175]: value = "task-1248382" [ 944.312215] env[63175]: _type = "Task" [ 944.312215] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.321911] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.367948] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 944.368386] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0b6bf590-1dfc-4cae-b67f-3ea1d01c1a60 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.380979] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 944.380979] env[63175]: value = "task-1248383" [ 944.380979] env[63175]: _type = "Task" [ 944.380979] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.390757] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248383, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.530764] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.620585] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.620875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.621117] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.621310] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.621496] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.624351] env[63175]: INFO nova.compute.manager [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Terminating instance [ 944.650576] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f428c6-f8cd-7e95-a821-afd30d75702b, 'name': SearchDatastore_Task, 'duration_secs': 0.030345} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.651472] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1216484-0fe8-4d74-8fba-0cb4639ef78a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.657522] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 944.657522] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c10e61-f7d8-4bae-9abf-f91ac505dcb9" [ 944.657522] env[63175]: _type = "Task" [ 944.657522] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.666590] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c10e61-f7d8-4bae-9abf-f91ac505dcb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.692766] env[63175]: INFO nova.compute.manager [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Unrescuing [ 944.693195] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.693388] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.693570] env[63175]: DEBUG nova.network.neutron [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.752641] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.752900] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.754556] env[63175]: INFO nova.compute.claims [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.824577] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248382, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.892214] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248383, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.031038] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.043144] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "089ce349-ac2d-4752-a41a-562b4acd2e73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.043387] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.128149] env[63175]: DEBUG nova.compute.manager [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 945.128379] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.129283] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872cfa9e-25e3-4c9b-9cdb-e2e4d28a2a1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.138023] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.138295] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c013494b-049d-462e-acf2-69b1b61a17d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.145099] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 945.145099] env[63175]: value = "task-1248384" [ 945.145099] env[63175]: _type = "Task" [ 945.145099] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.154118] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248384, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.167976] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c10e61-f7d8-4bae-9abf-f91ac505dcb9, 'name': SearchDatastore_Task, 'duration_secs': 0.011072} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.168294] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.168574] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 945.168853] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-002a3734-4462-4b75-8278-7e6c1f9d8e80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.176973] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 945.176973] env[63175]: value = "task-1248385" [ 945.176973] env[63175]: _type = "Task" [ 945.176973] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.186268] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.326940] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248382, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.395713] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248383, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.473550] env[63175]: DEBUG nova.network.neutron [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.533301] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.546043] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 945.655742] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248384, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.687181] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480078} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.687407] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.687974] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.687974] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2dad2e7-254e-42fd-bade-9cb670b09fcb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.695633] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 945.695633] env[63175]: value = "task-1248386" [ 945.695633] env[63175]: _type = "Task" [ 945.695633] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.711883] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.826836] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248382, 'name': PowerOffVM_Task, 'duration_secs': 1.442821} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.830135] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.830401] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.830948] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7e146dc-1c75-487b-a919-7d46dd6d3e81 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.897958] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248383, 'name': CloneVM_Task, 'duration_secs': 1.420118} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.898435] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Created linked-clone VM from snapshot [ 945.899431] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47407de3-6285-49e9-b641-8b9088861572 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.904127] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.904482] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.904815] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleting the datastore file [datastore1] cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.905912] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3904658f-2f79-4288-87d7-4076ff49ee78 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.915367] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Uploading image 8b773796-066f-4ad9-8133-ee8b138a0adc {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 945.923148] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 945.923148] env[63175]: value = "task-1248388" [ 945.923148] env[63175]: _type = "Task" [ 945.923148] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.932750] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.940231] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567d02dd-81fa-4af4-8c5e-48ad40495712 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.945251] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 945.945251] env[63175]: value = "vm-269075" [ 945.945251] env[63175]: _type = "VirtualMachine" [ 945.945251] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 945.945528] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f5444fe1-f8d8-4711-b45c-0645d8123c9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.953177] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af551bce-e4d8-4d3c-be24-67054a4e4bc2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.956629] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease: (returnval){ [ 945.956629] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524985c9-0079-47bc-dd97-22203b142a4a" [ 945.956629] env[63175]: _type = "HttpNfcLease" [ 945.956629] env[63175]: } obtained for exporting VM: (result){ [ 945.956629] env[63175]: value = "vm-269075" [ 945.956629] env[63175]: _type = "VirtualMachine" [ 945.956629] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 945.956898] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the lease: (returnval){ [ 945.956898] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524985c9-0079-47bc-dd97-22203b142a4a" [ 945.956898] env[63175]: _type = "HttpNfcLease" [ 945.956898] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 945.987141] env[63175]: DEBUG oslo_concurrency.lockutils [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.987885] env[63175]: DEBUG nova.objects.instance [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'flavor' on Instance uuid 0645fffe-9910-43af-af40-126592aefaff {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.990750] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3208511-c3f6-4411-bbc1-9cc85eb30c6b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.995055] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 945.995055] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524985c9-0079-47bc-dd97-22203b142a4a" [ 945.995055] env[63175]: _type = "HttpNfcLease" [ 945.995055] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 945.995712] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 945.995712] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524985c9-0079-47bc-dd97-22203b142a4a" [ 945.995712] env[63175]: _type = "HttpNfcLease" [ 945.995712] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 945.996607] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147fb57a-2124-4952-af60-bd199511e38f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.002613] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0017a8bb-741c-4b5c-9c24-df4d964d50a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.009872] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 946.010070] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 946.084514] env[63175]: DEBUG nova.compute.provider_tree [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.094103] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.098756] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.156172] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248384, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.206081] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073208} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.206419] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.207204] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b06925a-2b3b-4060-9193-a9e22e1b1be4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.229015] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.229589] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0dff122-6c4b-4d7f-85de-300d37372783 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.249258] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 946.249258] env[63175]: value = "task-1248390" [ 946.249258] env[63175]: _type = "Task" [ 946.249258] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.257723] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248390, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.328483] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0f723d62-2d74-48b5-b250-5cc992d4a2fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.434037] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.494453] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b829cc8-4433-4d64-a823-be7ef84eaf7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.519438] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.521838] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccbabc7c-6eea-420a-a4b6-8756e6cff3db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.530599] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 946.530599] env[63175]: value = "task-1248391" [ 946.530599] env[63175]: _type = "Task" [ 946.530599] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.541222] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248391, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.579052] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.589369] env[63175]: DEBUG nova.scheduler.client.report [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 946.657983] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248384, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.763465] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.934771] env[63175]: DEBUG oslo_vmware.api [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.743639} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.935240] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.935480] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.935744] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.935951] env[63175]: INFO nova.compute.manager [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Took 2.64 seconds to destroy the instance on the hypervisor. [ 946.936332] env[63175]: DEBUG oslo.service.loopingcall [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.936548] env[63175]: DEBUG nova.compute.manager [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 946.936649] env[63175]: DEBUG nova.network.neutron [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 947.041948] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248391, 'name': PowerOffVM_Task, 'duration_secs': 0.245716} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.042359] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.047658] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfiguring VM instance instance-0000004e to detach disk 2002 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 947.048437] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4293e8b-0863-44db-b298-da1a566dbb8c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.068867] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 947.068867] env[63175]: value = "task-1248392" [ 947.068867] env[63175]: _type = "Task" [ 947.068867] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.088822] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.089279] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248392, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.095634] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.096766] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 947.101483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.003s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.105169] env[63175]: INFO nova.compute.claims [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.159660] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248384, 'name': PowerOffVM_Task, 'duration_secs': 1.632239} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.160401] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.160732] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 947.161131] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8948f6f-be53-4036-818e-1d4d3eee2412 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.243634] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 947.244042] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 947.244345] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleting the datastore file [datastore2] aad1653a-43a7-4637-acd5-7dbf76ea1ccc {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.245425] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e74158a8-1ea6-4a4f-89b0-38684a54dc5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.256279] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 947.256279] env[63175]: value = "task-1248394" [ 947.256279] env[63175]: _type = "Task" [ 947.256279] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.264092] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248390, 'name': ReconfigVM_Task, 'duration_secs': 0.587099} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.264913] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.265701] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0506957-6790-49aa-bebb-3afa52e0d0f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.271633] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248394, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.278962] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 947.278962] env[63175]: value = "task-1248395" [ 947.278962] env[63175]: _type = "Task" [ 947.278962] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.290790] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248395, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.455908] env[63175]: DEBUG nova.compute.manager [req-403c7f14-6319-4c68-9f55-252928e1c7f7 req-1ab2e5bf-60b8-4fb1-ac0d-8a0405d28ceb service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Received event network-vif-deleted-71756e11-f67d-4268-a4b0-25b0a8cdefbf {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 947.456363] env[63175]: INFO nova.compute.manager [req-403c7f14-6319-4c68-9f55-252928e1c7f7 req-1ab2e5bf-60b8-4fb1-ac0d-8a0405d28ceb service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Neutron deleted interface 71756e11-f67d-4268-a4b0-25b0a8cdefbf; detaching it from the instance and deleting it from the info cache [ 947.456584] env[63175]: DEBUG nova.network.neutron [req-403c7f14-6319-4c68-9f55-252928e1c7f7 req-1ab2e5bf-60b8-4fb1-ac0d-8a0405d28ceb service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.582379] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248392, 'name': ReconfigVM_Task, 'duration_secs': 0.302498} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.587159] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfigured VM instance instance-0000004e to detach disk 2002 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 947.587159] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.587159] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.587159] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe45d119-1f94-45f5-bb12-70cd80b4bf46 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.597115] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 947.597115] env[63175]: value = "task-1248396" [ 947.597115] env[63175]: _type = "Task" [ 947.597115] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.605691] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248396, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.610462] env[63175]: DEBUG nova.compute.utils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.614496] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 947.614770] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.691992] env[63175]: DEBUG nova.policy [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '564ff9a1de5f43f4b3b05f3592bfa833', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54aee75c86d543a995a364ed78426ec2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.769011] env[63175]: DEBUG oslo_vmware.api [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248394, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23331} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.771277] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.771534] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 947.771766] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.772112] env[63175]: INFO nova.compute.manager [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Took 2.64 seconds to destroy the instance on the hypervisor. [ 947.772421] env[63175]: DEBUG oslo.service.loopingcall [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.772693] env[63175]: DEBUG nova.compute.manager [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 947.772806] env[63175]: DEBUG nova.network.neutron [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 947.789533] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248395, 'name': Rename_Task, 'duration_secs': 0.243914} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.790088] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.790088] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc755ee9-83fa-49be-a6b3-a71977aa982e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.798315] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 947.798315] env[63175]: value = "task-1248397" [ 947.798315] env[63175]: _type = "Task" [ 947.798315] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.807654] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248397, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.931588] env[63175]: DEBUG nova.network.neutron [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.963024] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7cd6240-b05c-45ff-a71a-e39333b56a21 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.975575] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1824a515-d8e2-47de-9862-79e9a2fcf8f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.011532] env[63175]: DEBUG nova.compute.manager [req-403c7f14-6319-4c68-9f55-252928e1c7f7 req-1ab2e5bf-60b8-4fb1-ac0d-8a0405d28ceb service nova] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Detach interface failed, port_id=71756e11-f67d-4268-a4b0-25b0a8cdefbf, reason: Instance cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 948.080187] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.106523] env[63175]: DEBUG oslo_vmware.api [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248396, 'name': PowerOnVM_Task, 'duration_secs': 0.418492} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.106523] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.106615] env[63175]: DEBUG nova.compute.manager [None req-41eb91cb-ff88-4f88-8745-5e786bd6e261 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.107479] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f68fb24-1a80-452d-8199-bb79aeea33a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.117339] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 948.148661] env[63175]: DEBUG nova.compute.manager [req-a96bb3b2-186d-4928-9bd0-2e49e4b486fc req-c96b00c1-0200-458b-9075-90536e17a738 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Received event network-vif-deleted-95854088-699b-4fec-a714-1b6959135fd2 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 948.148796] env[63175]: INFO nova.compute.manager [req-a96bb3b2-186d-4928-9bd0-2e49e4b486fc req-c96b00c1-0200-458b-9075-90536e17a738 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Neutron deleted interface 95854088-699b-4fec-a714-1b6959135fd2; detaching it from the instance and deleting it from the info cache [ 948.148995] env[63175]: DEBUG nova.network.neutron [req-a96bb3b2-186d-4928-9bd0-2e49e4b486fc req-c96b00c1-0200-458b-9075-90536e17a738 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.151427] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Successfully created port: 83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.312654] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248397, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.314798] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359836fa-b406-439c-94e3-c5e850ac21f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.323993] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d63c79-2741-440d-a466-058d88e0fe51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.358767] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a66d8e9-3822-4478-a066-4b76c7ec854f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.368037] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53289316-5f5f-4e8f-bf42-cc0f78e29020 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.383989] env[63175]: DEBUG nova.compute.provider_tree [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.434291] env[63175]: INFO nova.compute.manager [-] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Took 1.50 seconds to deallocate network for instance. [ 948.584234] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.614688] env[63175]: DEBUG nova.network.neutron [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.651810] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a0b088f-b689-4330-bcaa-6ceffe6f67ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.663452] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df06ccfa-8ca0-4268-84cf-d883ded6dd75 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.697537] env[63175]: DEBUG nova.compute.manager [req-a96bb3b2-186d-4928-9bd0-2e49e4b486fc req-c96b00c1-0200-458b-9075-90536e17a738 service nova] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Detach interface failed, port_id=95854088-699b-4fec-a714-1b6959135fd2, reason: Instance aad1653a-43a7-4637-acd5-7dbf76ea1ccc could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 948.816774] env[63175]: DEBUG oslo_vmware.api [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248397, 'name': PowerOnVM_Task, 'duration_secs': 0.568176} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.816774] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.817101] env[63175]: INFO nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Took 8.00 seconds to spawn the instance on the hypervisor. [ 948.817152] env[63175]: DEBUG nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.818303] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832fd931-972f-488c-92ec-b0bb1e8b4e40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.887044] env[63175]: DEBUG nova.scheduler.client.report [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 948.946635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.082961] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.117755] env[63175]: INFO nova.compute.manager [-] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Took 1.34 seconds to deallocate network for instance. [ 949.134081] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 949.163067] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.163234] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.163418] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.163779] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.164225] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.164328] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.164614] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.164882] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.165154] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.165379] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.165618] env[63175]: DEBUG nova.virt.hardware [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.167557] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a007d7-c871-4548-a9ee-d033babdd0a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.177928] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cfb8c5-4682-4ce8-bae2-8317395f3cfd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.341714] env[63175]: INFO nova.compute.manager [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Took 13.38 seconds to build instance. [ 949.392639] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.394432] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 949.397283] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.451s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.397283] env[63175]: DEBUG nova.objects.instance [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'resources' on Instance uuid cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.584831] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.611801] env[63175]: DEBUG nova.compute.manager [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 949.612166] env[63175]: DEBUG nova.compute.manager [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing instance network info cache due to event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 949.612430] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.612589] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.612764] env[63175]: DEBUG nova.network.neutron [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.626129] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.823954] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Successfully updated port: 83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.842889] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d60ad364-6b35-41c6-82cf-c5c72ad4e023 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.894s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.901572] env[63175]: DEBUG nova.compute.utils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.906726] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 949.907050] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.951565] env[63175]: DEBUG nova.policy [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c654f31fd4bc49dc8e985bd75a811ec2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74dd39e20aad4ddb874282973cea02bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.072954] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d9463e-8f8d-4aad-b96d-5d3d5dd8eb95 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.087618] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d0c210-9442-453e-8d9d-1d3cc68ed419 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.092021] env[63175]: DEBUG oslo_vmware.api [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248381, 'name': ReconfigVM_Task, 'duration_secs': 6.068932} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.092021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.092210] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Reconfigured VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 950.131888] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6559ed7f-2f42-4fe2-b82d-688f4ac60545 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.141584] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bb9229-2d5d-4ba8-a3d2-974b0a3f74b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.160796] env[63175]: DEBUG nova.compute.provider_tree [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.317114] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Successfully created port: 23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.333644] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.333873] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.334172] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.407900] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 950.450900] env[63175]: DEBUG nova.compute.manager [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Stashing vm_state: active {{(pid=63175) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 950.465637] env[63175]: DEBUG nova.compute.manager [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-deleted-b75f0d56-f89a-4754-a893-e06356ede92f {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 950.465835] env[63175]: INFO nova.compute.manager [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Neutron deleted interface b75f0d56-f89a-4754-a893-e06356ede92f; detaching it from the instance and deleting it from the info cache [ 950.466646] env[63175]: DEBUG nova.network.neutron [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "address": "fa:16:3e:37:9d:cd", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7e98f4d-1b", "ovs_interfaceid": "c7e98f4d-1b20-4610-ae0d-74b0c13618ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.554647] env[63175]: DEBUG nova.network.neutron [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updated VIF entry in instance network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.555145] env[63175]: DEBUG nova.network.neutron [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.664434] env[63175]: DEBUG nova.scheduler.client.report [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 950.873557] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.972105] env[63175]: DEBUG oslo_concurrency.lockutils [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.972309] env[63175]: DEBUG oslo_concurrency.lockutils [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] Acquired lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.973600] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa25daa-9b4b-47b0-9ba9-c5480a9feb89 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.981758] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.998162] env[63175]: DEBUG oslo_concurrency.lockutils [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] Releasing lock "a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.998521] env[63175]: WARNING nova.compute.manager [req-ee958543-211f-4ba1-8c5e-dde21d044b20 req-bcbcb16c-92d8-4075-b040-1d2b5be23c77 service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Detach interface failed, port_id=b75f0d56-f89a-4754-a893-e06356ede92f, reason: No device with interface-id b75f0d56-f89a-4754-a893-e06356ede92f exists on VM: nova.exception.NotFound: No device with interface-id b75f0d56-f89a-4754-a893-e06356ede92f exists on VM [ 951.058632] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d877204-b461-4488-b7a2-d97f43c1248c req-38e88a59-a9fc-4426-84cd-8c7af98c703d service nova] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.077656] env[63175]: DEBUG nova.network.neutron [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updating instance_info_cache with network_info: [{"id": "83181b9f-233b-471b-bb93-9c8cafed721b", "address": "fa:16:3e:31:05:9e", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83181b9f-23", "ovs_interfaceid": "83181b9f-233b-471b-bb93-9c8cafed721b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.169229] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.772s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.172573] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.547s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.172699] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.176841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.193s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.191286] env[63175]: INFO nova.scheduler.client.report [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleted allocations for instance cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3 [ 951.197038] env[63175]: INFO nova.scheduler.client.report [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocations for instance aad1653a-43a7-4637-acd5-7dbf76ea1ccc [ 951.420745] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 951.443123] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.443446] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.443634] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.443873] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.444147] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.444329] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.444590] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.444792] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.445014] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.445221] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.445405] env[63175]: DEBUG nova.virt.hardware [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.445888] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.446064] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.446331] env[63175]: DEBUG nova.network.neutron [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.448095] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b796252a-aae0-414f-9495-76858d22084e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.459048] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfd4f73-d270-4c23-a9ba-195055fb82c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.580875] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.581216] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Instance network_info: |[{"id": "83181b9f-233b-471b-bb93-9c8cafed721b", "address": "fa:16:3e:31:05:9e", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83181b9f-23", "ovs_interfaceid": "83181b9f-233b-471b-bb93-9c8cafed721b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 951.582033] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:05:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83181b9f-233b-471b-bb93-9c8cafed721b', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.589709] env[63175]: DEBUG oslo.service.loopingcall [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.590476] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.590941] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5768c4d0-7f38-40c6-a3df-009dd884e04e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.612854] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.612854] env[63175]: value = "task-1248398" [ 951.612854] env[63175]: _type = "Task" [ 951.612854] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.622210] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248398, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.681582] env[63175]: INFO nova.compute.claims [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.705090] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b6385a70-3e31-425e-ac87-09bc859b6d4d tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.924s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.706572] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d00d8cb5-0515-4824-a7fe-1a919d53e67f tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "aad1653a-43a7-4637-acd5-7dbf76ea1ccc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.086s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.718582] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 951.718782] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing instance network info cache due to event network-changed-a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 951.719018] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Acquiring lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.719180] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Acquired lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.719345] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Refreshing network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.971170] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Successfully updated port: 23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.075493] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.075780] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.076182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "a4304579-f829-433e-a878-1050fac08c2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.076389] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.076581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.078683] env[63175]: INFO nova.compute.manager [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Terminating instance [ 952.126376] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248398, 'name': CreateVM_Task, 'duration_secs': 0.360105} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.126376] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.126981] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.127222] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.127573] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.127838] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e4572a0-6ffa-40fe-abd2-cee919925c7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.133158] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 952.133158] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52872c93-3d18-5838-4db7-aa6a48914f6f" [ 952.133158] env[63175]: _type = "Task" [ 952.133158] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.141301] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52872c93-3d18-5838-4db7-aa6a48914f6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.188710] env[63175]: INFO nova.compute.resource_tracker [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating resource usage from migration dbd8eafe-cfcc-4812-b66e-45973401f12f [ 952.283892] env[63175]: INFO nova.network.neutron [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Port c7e98f4d-1b20-4610-ae0d-74b0c13618ba from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 952.284377] env[63175]: DEBUG nova.network.neutron [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [{"id": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "address": "fa:16:3e:66:3e:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cd631a8-97", "ovs_interfaceid": "3cd631a8-97d6-4ff8-87c6-ab78b6d39c51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.346996] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dbe0f1-7e78-4849-8bb8-66f822400cc4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.356044] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c86359-ab1b-4a40-ade7-7ff15d40117f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.390201] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117b7420-712a-44d0-8279-e7f6dc6e4402 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.398942] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86ed979-4274-4af9-a347-d81e6e06abfc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.413807] env[63175]: DEBUG nova.compute.provider_tree [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.476382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.476382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.476382] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.504374] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updated VIF entry in instance network info cache for port a16e4e4f-df78-450a-bf4d-d43daac38791. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.504749] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [{"id": "a16e4e4f-df78-450a-bf4d-d43daac38791", "address": "fa:16:3e:8c:a2:ac", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16e4e4f-df", "ovs_interfaceid": "a16e4e4f-df78-450a-bf4d-d43daac38791", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.523155] env[63175]: DEBUG nova.compute.manager [req-a1d0c3ec-963a-4a64-a520-2a718c953807 req-3bfe737d-f97b-4064-991c-fa74fb22344a service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-deleted-c7e98f4d-1b20-4610-ae0d-74b0c13618ba {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 952.584530] env[63175]: DEBUG nova.compute.manager [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 952.584762] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.585727] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1b7b63-2fbb-43a1-9f5e-6e48b510ee14 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.594986] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.595311] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5304fba3-2c2c-428a-9f42-9d15c2bb8382 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.602458] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 952.602458] env[63175]: value = "task-1248399" [ 952.602458] env[63175]: _type = "Task" [ 952.602458] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.613223] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.643815] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52872c93-3d18-5838-4db7-aa6a48914f6f, 'name': SearchDatastore_Task, 'duration_secs': 0.016482} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.644248] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.644567] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.644842] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.645046] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.645282] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.645584] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcbfde3b-f06f-4e54-9d58-44a147756a59 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.657663] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.657994] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 952.658603] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b32f9ef-cf65-47c5-9a18-57134965dcce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.664381] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 952.664381] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52445274-cf31-a6b8-a1f1-970f44185d0d" [ 952.664381] env[63175]: _type = "Task" [ 952.664381] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.672345] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52445274-cf31-a6b8-a1f1-970f44185d0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.787635] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-a4304579-f829-433e-a878-1050fac08c2e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.917568] env[63175]: DEBUG nova.scheduler.client.report [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 953.008473] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Releasing lock "refresh_cache-0645fffe-9910-43af-af40-126592aefaff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.009902] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Received event network-vif-plugged-83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 953.010227] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.010508] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.010693] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.010882] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] No waiting events found dispatching network-vif-plugged-83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 953.011063] env[63175]: WARNING nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Received unexpected event network-vif-plugged-83181b9f-233b-471b-bb93-9c8cafed721b for instance with vm_state building and task_state spawning. [ 953.011244] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Received event network-changed-83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 953.011406] env[63175]: DEBUG nova.compute.manager [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Refreshing instance network info cache due to event network-changed-83181b9f-233b-471b-bb93-9c8cafed721b. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 953.011597] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Acquiring lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.011739] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Acquired lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.011904] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Refreshing network info cache for port 83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.022692] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 953.115286] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248399, 'name': PowerOffVM_Task, 'duration_secs': 0.301266} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.115610] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.115847] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.116138] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d6e5dae-3f34-42a2-ab35-3d2bc8d2d2d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.176302] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52445274-cf31-a6b8-a1f1-970f44185d0d, 'name': SearchDatastore_Task, 'duration_secs': 0.023278} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.177803] env[63175]: DEBUG nova.network.neutron [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Updating instance_info_cache with network_info: [{"id": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "address": "fa:16:3e:15:c3:7a", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23c61658-90", "ovs_interfaceid": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.178838] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12919c59-9e1d-4cb9-a178-eace8ca09f2b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.185721] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 953.185721] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528c3709-48cd-0dbe-1565-9bbe1aacc4a5" [ 953.185721] env[63175]: _type = "Task" [ 953.185721] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.195641] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528c3709-48cd-0dbe-1565-9bbe1aacc4a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.242494] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.242703] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.242892] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleting the datastore file [datastore1] a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.245023] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eab99966-fcdd-4fc1-b76e-395c2c894241 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.251445] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 953.251445] env[63175]: value = "task-1248401" [ 953.251445] env[63175]: _type = "Task" [ 953.251445] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.261821] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.293204] env[63175]: DEBUG oslo_concurrency.lockutils [None req-223e68aa-a8ae-46f2-971a-df59a95a63ea tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-a4304579-f829-433e-a878-1050fac08c2e-b75f0d56-f89a-4754-a893-e06356ede92f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.350s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.423370] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.248s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.423581] env[63175]: INFO nova.compute.manager [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Migrating [ 953.683026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.683404] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Instance network_info: |[{"id": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "address": "fa:16:3e:15:c3:7a", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23c61658-90", "ovs_interfaceid": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 953.683881] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:c3:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cde23701-02ca-4cb4-b5a6-d321f8ac9660', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23c61658-90f5-4d6c-9db5-24b05bd5e37c', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.694083] env[63175]: DEBUG oslo.service.loopingcall [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.694421] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.698161] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ae5ebb9-c1b7-447b-b5f6-6678124b6847 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.729817] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528c3709-48cd-0dbe-1565-9bbe1aacc4a5, 'name': SearchDatastore_Task, 'duration_secs': 0.019066} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.731297] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.731564] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5/2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.731806] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.731806] env[63175]: value = "task-1248402" [ 953.731806] env[63175]: _type = "Task" [ 953.731806] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.732053] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0f0eca2-b741-435e-9aa0-feb3190f4522 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.742592] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248402, 'name': CreateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.743876] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 953.743876] env[63175]: value = "task-1248403" [ 953.743876] env[63175]: _type = "Task" [ 953.743876] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.752187] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248403, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.760485] env[63175]: DEBUG oslo_vmware.api [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.420013} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.760731] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.760921] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.761120] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.761304] env[63175]: INFO nova.compute.manager [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: a4304579-f829-433e-a878-1050fac08c2e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 953.761546] env[63175]: DEBUG oslo.service.loopingcall [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.761742] env[63175]: DEBUG nova.compute.manager [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 953.761826] env[63175]: DEBUG nova.network.neutron [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.814068] env[63175]: DEBUG nova.compute.manager [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Received event network-vif-plugged-23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 953.814358] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Acquiring lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.814529] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.814778] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.814901] env[63175]: DEBUG nova.compute.manager [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] No waiting events found dispatching network-vif-plugged-23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 953.815355] env[63175]: WARNING nova.compute.manager [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Received unexpected event network-vif-plugged-23c61658-90f5-4d6c-9db5-24b05bd5e37c for instance with vm_state building and task_state spawning. [ 953.815593] env[63175]: DEBUG nova.compute.manager [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Received event network-changed-23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 953.815763] env[63175]: DEBUG nova.compute.manager [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Refreshing instance network info cache due to event network-changed-23c61658-90f5-4d6c-9db5-24b05bd5e37c. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 953.815956] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Acquiring lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.816118] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Acquired lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.816285] env[63175]: DEBUG nova.network.neutron [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Refreshing network info cache for port 23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.940077] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.940378] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.940660] env[63175]: DEBUG nova.network.neutron [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.986866] env[63175]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port c7e98f4d-1b20-4610-ae0d-74b0c13618ba could not be found.", "detail": ""}} {{(pid=63175) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 953.987165] env[63175]: DEBUG nova.network.neutron [-] Unable to show port c7e98f4d-1b20-4610-ae0d-74b0c13618ba as it no longer exists. {{(pid=63175) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 954.246518] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248402, 'name': CreateVM_Task, 'duration_secs': 0.375081} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.249847] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.250627] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.250807] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.251169] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.252157] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updated VIF entry in instance network info cache for port 83181b9f-233b-471b-bb93-9c8cafed721b. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.252521] env[63175]: DEBUG nova.network.neutron [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updating instance_info_cache with network_info: [{"id": "83181b9f-233b-471b-bb93-9c8cafed721b", "address": "fa:16:3e:31:05:9e", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83181b9f-23", "ovs_interfaceid": "83181b9f-233b-471b-bb93-9c8cafed721b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.254279] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0402f91a-87e5-41d1-bb6e-ab3284332d08 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.261752] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248403, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.263504] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 954.263504] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5236ab86-f974-81be-5f5b-0183e160f934" [ 954.263504] env[63175]: _type = "Task" [ 954.263504] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.275801] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5236ab86-f974-81be-5f5b-0183e160f934, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.676757] env[63175]: DEBUG nova.network.neutron [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Updated VIF entry in instance network info cache for port 23c61658-90f5-4d6c-9db5-24b05bd5e37c. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.680025] env[63175]: DEBUG nova.network.neutron [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Updating instance_info_cache with network_info: [{"id": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "address": "fa:16:3e:15:c3:7a", "network": {"id": "391fea5c-91a1-4dc9-b35f-01e4c72ff9a4", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1853804315-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "74dd39e20aad4ddb874282973cea02bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cde23701-02ca-4cb4-b5a6-d321f8ac9660", "external-id": "nsx-vlan-transportzone-586", "segmentation_id": 586, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23c61658-90", "ovs_interfaceid": "23c61658-90f5-4d6c-9db5-24b05bd5e37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.726018] env[63175]: DEBUG nova.network.neutron [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.759117] env[63175]: DEBUG oslo_concurrency.lockutils [req-dd5008bd-4522-42ed-b2a0-e1fad6efa34a req-05d1854f-eaf3-4e9c-9ce1-78f8e51c0833 service nova] Releasing lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.759825] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248403, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702044} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.760373] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5/2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.760623] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.760920] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-215a35cb-18dd-42fd-90cd-fa1ccca33ef4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.769756] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 954.769756] env[63175]: value = "task-1248404" [ 954.769756] env[63175]: _type = "Task" [ 954.769756] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.778480] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5236ab86-f974-81be-5f5b-0183e160f934, 'name': SearchDatastore_Task, 'duration_secs': 0.054554} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.781197] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.781420] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.781681] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.781867] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.782136] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.785232] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-639c58ec-6bc0-49a2-9316-3c3c9180d270 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.787320] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248404, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.794929] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.795176] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.796124] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-380e1f30-80ac-419a-879b-3eda6297be7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.801701] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 954.801701] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f1bd86-3e20-423d-b8b1-94c907a6c6d1" [ 954.801701] env[63175]: _type = "Task" [ 954.801701] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.811293] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f1bd86-3e20-423d-b8b1-94c907a6c6d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.853823] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 954.855119] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81881920-6acf-4319-a5ab-fa01435e2367 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.861520] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 954.861701] env[63175]: ERROR oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk due to incomplete transfer. [ 954.861938] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9eb9a35d-1955-4c37-86cd-a79cb3e886b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.869103] env[63175]: DEBUG oslo_vmware.rw_handles [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fddc3a-7b90-7ba5-d187-4c152e65103e/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 954.869310] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Uploaded image 8b773796-066f-4ad9-8133-ee8b138a0adc to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 954.871564] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 954.871830] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f94a1df4-a598-4721-b6d3-1ae32b83a057 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.881766] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 954.881766] env[63175]: value = "task-1248405" [ 954.881766] env[63175]: _type = "Task" [ 954.881766] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.895430] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248405, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.981691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.981874] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.180229] env[63175]: DEBUG oslo_concurrency.lockutils [req-d82b93dc-723d-4394-8470-d710515c2591 req-60261176-eb92-49f4-b1e6-3dccbef364bc service nova] Releasing lock "refresh_cache-089ce349-ac2d-4752-a41a-562b4acd2e73" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.228260] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.242088] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.242831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.242831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.242831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.243012] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.246170] env[63175]: INFO nova.compute.manager [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Terminating instance [ 955.247548] env[63175]: DEBUG nova.network.neutron [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.280184] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248404, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07466} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.280466] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.281281] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998d7c03-0889-449c-8bc6-6bdf399feb4b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.303929] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5/2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.304399] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f35b5b0-42b0-4284-b107-3c1dd6e79358 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.330187] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f1bd86-3e20-423d-b8b1-94c907a6c6d1, 'name': SearchDatastore_Task, 'duration_secs': 0.017072} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.332252] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 955.332252] env[63175]: value = "task-1248406" [ 955.332252] env[63175]: _type = "Task" [ 955.332252] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.332502] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e0c47a-3654-433e-a7d1-98c78e5dd2a0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.341791] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 955.341791] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290e73d-190e-a494-31f0-89605006e88c" [ 955.341791] env[63175]: _type = "Task" [ 955.341791] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.345201] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.353063] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290e73d-190e-a494-31f0-89605006e88c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.392015] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248405, 'name': Destroy_Task, 'duration_secs': 0.477683} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.392401] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Destroyed the VM [ 955.392674] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 955.392947] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c189ec54-4ae4-46c9-9d82-e6d912c4ab02 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.400244] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 955.400244] env[63175]: value = "task-1248407" [ 955.400244] env[63175]: _type = "Task" [ 955.400244] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.410653] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248407, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.485673] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 955.749783] env[63175]: DEBUG nova.compute.manager [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 955.750053] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.750474] env[63175]: INFO nova.compute.manager [-] [instance: a4304579-f829-433e-a878-1050fac08c2e] Took 1.99 seconds to deallocate network for instance. [ 955.751418] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29ae653-c00d-4f72-a76b-08574345b8cc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.762092] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.762383] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eba0589-6ae1-4aa2-a993-8c493fb6cd88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.769711] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 955.769711] env[63175]: value = "task-1248408" [ 955.769711] env[63175]: _type = "Task" [ 955.769711] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.779578] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.845302] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248406, 'name': ReconfigVM_Task, 'duration_secs': 0.370905} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.847081] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5/2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.848940] env[63175]: DEBUG nova.compute.manager [req-e2401f05-06ea-4a6b-a324-9fc77be5659b req-c74b535e-4dc9-4495-9af7-c390e9c08edd service nova] [instance: a4304579-f829-433e-a878-1050fac08c2e] Received event network-vif-deleted-3cd631a8-97d6-4ff8-87c6-ab78b6d39c51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 955.852826] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-477fd487-c483-4277-b84e-9cf5f1050855 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.862253] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5290e73d-190e-a494-31f0-89605006e88c, 'name': SearchDatastore_Task, 'duration_secs': 0.034514} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.863686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.864077] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 089ce349-ac2d-4752-a41a-562b4acd2e73/089ce349-ac2d-4752-a41a-562b4acd2e73.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.864352] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 955.864352] env[63175]: value = "task-1248409" [ 955.864352] env[63175]: _type = "Task" [ 955.864352] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.864563] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47af4b00-e9fe-44e8-912d-555bab9988c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.877494] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248409, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.878734] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 955.878734] env[63175]: value = "task-1248410" [ 955.878734] env[63175]: _type = "Task" [ 955.878734] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.888667] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248410, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.911244] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248407, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.009093] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.009493] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.012211] env[63175]: INFO nova.compute.claims [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.261674] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.282103] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248408, 'name': PowerOffVM_Task, 'duration_secs': 0.404098} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.282424] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.282686] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.283042] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f06bd03a-2776-4220-a093-80e8675017d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.374870] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.375179] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.375465] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleting the datastore file [datastore2] 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.375850] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edca9cbc-5128-4294-a2eb-f4cb3b9b07e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.382505] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248409, 'name': Rename_Task, 'duration_secs': 0.173938} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.387355] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.387745] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for the task: (returnval){ [ 956.387745] env[63175]: value = "task-1248412" [ 956.387745] env[63175]: _type = "Task" [ 956.387745] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.387991] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a385e19d-3a6f-4435-b3f2-51ac32d73b3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.398967] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248410, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.410203] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.410673] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 956.410673] env[63175]: value = "task-1248413" [ 956.410673] env[63175]: _type = "Task" [ 956.410673] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.417972] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248407, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.423719] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.446433] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.446710] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.748035] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6965e906-0342-4e50-8466-55d3aa4470b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.766243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 0 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.892947] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248410, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529544} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.896419] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 089ce349-ac2d-4752-a41a-562b4acd2e73/089ce349-ac2d-4752-a41a-562b4acd2e73.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.896614] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.896791] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-394c398d-7827-46df-85eb-1e2d35a785b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.905520] env[63175]: DEBUG oslo_vmware.api [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Task: {'id': task-1248412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211878} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.910216] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.910725] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.910873] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.911421] env[63175]: INFO nova.compute.manager [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 956.911753] env[63175]: DEBUG oslo.service.loopingcall [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.912047] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 956.912047] env[63175]: value = "task-1248414" [ 956.912047] env[63175]: _type = "Task" [ 956.912047] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.912391] env[63175]: DEBUG nova.compute.manager [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 956.912494] env[63175]: DEBUG nova.network.neutron [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.925826] env[63175]: DEBUG oslo_vmware.api [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248407, 'name': RemoveSnapshot_Task, 'duration_secs': 1.254204} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.926169] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 956.926424] env[63175]: INFO nova.compute.manager [None req-38846840-d818-4979-9ed8-eb1d22c45f79 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Took 15.14 seconds to snapshot the instance on the hypervisor. [ 956.934634] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.934897] env[63175]: DEBUG oslo_vmware.api [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248413, 'name': PowerOnVM_Task, 'duration_secs': 0.515058} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.935526] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.935638] env[63175]: INFO nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Took 7.80 seconds to spawn the instance on the hypervisor. [ 956.935822] env[63175]: DEBUG nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 956.936629] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0336fadd-eb9c-46fc-8e35-98cef618b105 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.948948] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 957.193032] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be6b098-ee45-40f9-8257-23c7a0586c01 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.201200] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0073b27f-256c-49c7-95a8-0fffbf354d38 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.231756] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f916c907-3485-4ab5-bf6b-3f8fd2a4f121 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.241217] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3797ad90-2974-47f7-9460-ba2f37b68f61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.255889] env[63175]: DEBUG nova.compute.provider_tree [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.271949] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.272484] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5813317d-d422-41a3-957e-75d8c81367f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.280535] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 957.280535] env[63175]: value = "task-1248415" [ 957.280535] env[63175]: _type = "Task" [ 957.280535] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.289640] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248415, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.425487] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073293} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.425795] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.426662] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9b70bb-d7bf-4459-a63d-9e37c229bc65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.453841] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 089ce349-ac2d-4752-a41a-562b4acd2e73/089ce349-ac2d-4752-a41a-562b4acd2e73.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.460771] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9852b4f3-6b83-4b1b-bdab-b6c2b535a0ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.482123] env[63175]: INFO nova.compute.manager [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Took 12.74 seconds to build instance. [ 957.489293] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 957.489293] env[63175]: value = "task-1248416" [ 957.489293] env[63175]: _type = "Task" [ 957.489293] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.495348] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.499861] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248416, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.758664] env[63175]: DEBUG nova.scheduler.client.report [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 957.790983] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248415, 'name': PowerOffVM_Task, 'duration_secs': 0.359988} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.791285] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.791481] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 17 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.856753] env[63175]: DEBUG nova.network.neutron [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.872302] env[63175]: DEBUG nova.compute.manager [req-39b28588-51bf-42b3-aa16-ff9ce584a061 req-f5c957fd-6dbe-4838-b54a-cff03c4c7d31 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Received event network-vif-deleted-6a69f7df-439f-4044-8394-305f9f3a93bb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 957.872455] env[63175]: INFO nova.compute.manager [req-39b28588-51bf-42b3-aa16-ff9ce584a061 req-f5c957fd-6dbe-4838-b54a-cff03c4c7d31 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Neutron deleted interface 6a69f7df-439f-4044-8394-305f9f3a93bb; detaching it from the instance and deleting it from the info cache [ 957.872627] env[63175]: DEBUG nova.network.neutron [req-39b28588-51bf-42b3-aa16-ff9ce584a061 req-f5c957fd-6dbe-4838-b54a-cff03c4c7d31 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.984794] env[63175]: DEBUG oslo_concurrency.lockutils [None req-51134e60-55a4-4b01-bf61-2b9207c8e46c tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.254s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.999814] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248416, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.266467] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.266467] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 958.269920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.007s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.269920] env[63175]: DEBUG nova.objects.instance [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'resources' on Instance uuid a4304579-f829-433e-a878-1050fac08c2e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.299306] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:07Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.301118] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.301118] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.301118] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.301118] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.301324] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.301803] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.302384] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.302551] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.302671] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.302965] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.308843] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f57561f2-5e4f-4ed0-ad6f-00fe3c602e82 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.326335] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 958.326335] env[63175]: value = "task-1248417" [ 958.326335] env[63175]: _type = "Task" [ 958.326335] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.335227] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248417, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.359824] env[63175]: INFO nova.compute.manager [-] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Took 1.45 seconds to deallocate network for instance. [ 958.376183] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c114729e-6ed3-41d5-ac9f-09dc07853554 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.386021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19297ab-236e-45b8-95fe-fa4d61ff58b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.414638] env[63175]: DEBUG nova.compute.manager [req-39b28588-51bf-42b3-aa16-ff9ce584a061 req-f5c957fd-6dbe-4838-b54a-cff03c4c7d31 service nova] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Detach interface failed, port_id=6a69f7df-439f-4044-8394-305f9f3a93bb, reason: Instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 958.502385] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248416, 'name': ReconfigVM_Task, 'duration_secs': 0.668227} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.502678] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 089ce349-ac2d-4752-a41a-562b4acd2e73/089ce349-ac2d-4752-a41a-562b4acd2e73.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.503326] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5ba3933-5b04-4339-8d3a-41a056d495c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.510363] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 958.510363] env[63175]: value = "task-1248418" [ 958.510363] env[63175]: _type = "Task" [ 958.510363] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.522226] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248418, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.776310] env[63175]: DEBUG nova.compute.utils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.780478] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 958.780648] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 958.822807] env[63175]: DEBUG nova.policy [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14a8b3d2f6574e37b3125aba432bc56a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e8826d0f0384b56aa862756787ad017', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.839087] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248417, 'name': ReconfigVM_Task, 'duration_secs': 0.256476} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.839452] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 33 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.865660] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.959157] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9536543-7ca7-4ccc-9d91-c0ac1ad8d734 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.970380] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b34fb6-20b9-47b9-8d1b-7f82a872ff68 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.005023] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2821b4a-2ce8-4f6b-ba11-1bdcfe9249a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.015066] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7581d4ce-dafa-4751-a96a-d23c41d5f007 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.031352] env[63175]: DEBUG nova.compute.provider_tree [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.035180] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248418, 'name': Rename_Task, 'duration_secs': 0.160525} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.035416] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.035683] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9638c209-6868-4902-95db-fb2808744fa8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.044715] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 959.044715] env[63175]: value = "task-1248419" [ 959.044715] env[63175]: _type = "Task" [ 959.044715] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.054058] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.157062] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Successfully created port: 9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.281584] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.349137] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.349651] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.349975] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.350298] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.350610] env[63175]: DEBUG nova.virt.hardware [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.357413] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 959.357413] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2768b3b-657e-4a2d-be5f-3447700a2df3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.381662] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 959.381662] env[63175]: value = "task-1248420" [ 959.381662] env[63175]: _type = "Task" [ 959.381662] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.391805] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.540811] env[63175]: DEBUG nova.scheduler.client.report [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 959.556250] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248419, 'name': PowerOnVM_Task} progress is 95%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.893391] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248420, 'name': ReconfigVM_Task, 'duration_secs': 0.180262} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.893554] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 959.894452] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882ee96c-51e6-424b-9777-a129a08ed49f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.900100] env[63175]: DEBUG nova.compute.manager [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Received event network-changed-83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 959.900411] env[63175]: DEBUG nova.compute.manager [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Refreshing instance network info cache due to event network-changed-83181b9f-233b-471b-bb93-9c8cafed721b. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 959.900647] env[63175]: DEBUG oslo_concurrency.lockutils [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] Acquiring lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.900845] env[63175]: DEBUG oslo_concurrency.lockutils [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] Acquired lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.901085] env[63175]: DEBUG nova.network.neutron [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Refreshing network info cache for port 83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.925894] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.927153] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9c2c83d-7d09-433d-996e-8b542ca44751 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.947405] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 959.947405] env[63175]: value = "task-1248421" [ 959.947405] env[63175]: _type = "Task" [ 959.947405] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.956429] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248421, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.046468] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.777s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.049383] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.554s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.051046] env[63175]: INFO nova.compute.claims [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.064383] env[63175]: DEBUG oslo_vmware.api [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248419, 'name': PowerOnVM_Task, 'duration_secs': 0.558618} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.064711] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.064952] env[63175]: INFO nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Took 8.64 seconds to spawn the instance on the hypervisor. [ 960.065185] env[63175]: DEBUG nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 960.066072] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d707d5c0-7cdf-4270-9182-fd91ff1988ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.076201] env[63175]: INFO nova.scheduler.client.report [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted allocations for instance a4304579-f829-433e-a878-1050fac08c2e [ 960.292103] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 960.315274] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.315551] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.315716] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.315905] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.316068] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.316229] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.316440] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.316600] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.316770] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.316934] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.317125] env[63175]: DEBUG nova.virt.hardware [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.317961] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c03b648-4add-48aa-a00a-ccb6fdeb9cf3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.326154] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452968a5-8300-4de3-b579-1a7cce7d808b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.458501] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248421, 'name': ReconfigVM_Task, 'duration_secs': 0.281848} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.458857] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df/0633c8e9-1c18-4047-ae5d-012d715e03df.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.459154] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 50 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 960.577134] env[63175]: DEBUG nova.compute.manager [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Received event network-vif-plugged-9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.577399] env[63175]: DEBUG oslo_concurrency.lockutils [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] Acquiring lock "572a3821-7436-487a-a053-3819411de57e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.577613] env[63175]: DEBUG oslo_concurrency.lockutils [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] Lock "572a3821-7436-487a-a053-3819411de57e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.577789] env[63175]: DEBUG oslo_concurrency.lockutils [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] Lock "572a3821-7436-487a-a053-3819411de57e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.577968] env[63175]: DEBUG nova.compute.manager [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] No waiting events found dispatching network-vif-plugged-9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 960.578295] env[63175]: WARNING nova.compute.manager [req-abc3d20c-7cf2-41c7-b446-043c888d5fbd req-434ac17b-4973-419a-bb64-4d7f4daa5aa3 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Received unexpected event network-vif-plugged-9f3d24b2-7d86-4548-a817-f8587b820d00 for instance with vm_state building and task_state spawning. [ 960.590893] env[63175]: DEBUG oslo_concurrency.lockutils [None req-55f47c13-5e29-4e63-827d-bf3e2dbfb8b3 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "a4304579-f829-433e-a878-1050fac08c2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.515s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.593144] env[63175]: INFO nova.compute.manager [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Took 14.53 seconds to build instance. [ 960.634625] env[63175]: DEBUG nova.network.neutron [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updated VIF entry in instance network info cache for port 83181b9f-233b-471b-bb93-9c8cafed721b. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 960.635043] env[63175]: DEBUG nova.network.neutron [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updating instance_info_cache with network_info: [{"id": "83181b9f-233b-471b-bb93-9c8cafed721b", "address": "fa:16:3e:31:05:9e", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83181b9f-23", "ovs_interfaceid": "83181b9f-233b-471b-bb93-9c8cafed721b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.966092] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788dd68f-2c36-4a2c-a66d-0d64933dd965 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.986369] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e63657-daf0-499a-ba58-168bad0f5bfc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.005886] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 67 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.097554] env[63175]: DEBUG oslo_concurrency.lockutils [None req-570b0dc0-8b1a-45bb-8325-176cf801f962 tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.054s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.137542] env[63175]: DEBUG oslo_concurrency.lockutils [req-03e8a5e0-ae7d-4aba-9599-eb636249ad7e req-e4d78e30-5b3a-486f-aeed-2223abbbde63 service nova] Releasing lock "refresh_cache-2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.211183] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Successfully updated port: 9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.242334] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24483212-7b1e-4d3f-8cc3-92793cfd143c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.250537] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85bd633-e272-4adc-b127-a8860e32640c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.282079] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8226503-79ff-47e0-9fe8-84621b469695 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.290844] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a4c662-d50b-491a-b8e9-e4b4695f172e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.305307] env[63175]: DEBUG nova.compute.provider_tree [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.547774] env[63175]: DEBUG nova.network.neutron [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Port 4621f750-c3ff-4289-bf4f-9ad952a25f05 binding to destination host cpu-1 is already ACTIVE {{(pid=63175) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 961.713800] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.714122] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.714699] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.808849] env[63175]: DEBUG nova.scheduler.client.report [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 961.969364] env[63175]: DEBUG nova.compute.manager [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Received event network-changed-9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 961.969574] env[63175]: DEBUG nova.compute.manager [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Refreshing instance network info cache due to event network-changed-9f3d24b2-7d86-4548-a817-f8587b820d00. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 961.969765] env[63175]: DEBUG oslo_concurrency.lockutils [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] Acquiring lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.250387] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 962.314684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.315221] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 962.317796] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.452s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.317998] env[63175]: DEBUG nova.objects.instance [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lazy-loading 'resources' on Instance uuid 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.440901] env[63175]: DEBUG nova.network.neutron [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating instance_info_cache with network_info: [{"id": "9f3d24b2-7d86-4548-a817-f8587b820d00", "address": "fa:16:3e:ee:45:0c", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f3d24b2-7d", "ovs_interfaceid": "9f3d24b2-7d86-4548-a817-f8587b820d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.527114] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "089ce349-ac2d-4752-a41a-562b4acd2e73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.527428] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.527658] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.527843] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.528032] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.530332] env[63175]: INFO nova.compute.manager [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Terminating instance [ 962.541672] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.541893] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.568738] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.568915] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.569114] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.821612] env[63175]: DEBUG nova.compute.utils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.826042] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 962.826545] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 962.867825] env[63175]: DEBUG nova.policy [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e43774590162406a981cf0d5e20f5a1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83fa8f008a48484195c88ed76ceece0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.944454] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.945126] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Instance network_info: |[{"id": "9f3d24b2-7d86-4548-a817-f8587b820d00", "address": "fa:16:3e:ee:45:0c", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f3d24b2-7d", "ovs_interfaceid": "9f3d24b2-7d86-4548-a817-f8587b820d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 962.946026] env[63175]: DEBUG oslo_concurrency.lockutils [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] Acquired lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.946026] env[63175]: DEBUG nova.network.neutron [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Refreshing network info cache for port 9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.946991] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:45:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '777870ab-362f-4a17-9c1c-8d9cc26cd4ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f3d24b2-7d86-4548-a817-f8587b820d00', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.955793] env[63175]: DEBUG oslo.service.loopingcall [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.959417] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 572a3821-7436-487a-a053-3819411de57e] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.962105] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3557428e-8b81-4619-bff1-cd1826c114b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.984144] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.984144] env[63175]: value = "task-1248422" [ 962.984144] env[63175]: _type = "Task" [ 962.984144] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.994401] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248422, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.008120] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bd0cea-0d80-4414-b007-559f072e06e4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.016876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a60ff9-89cd-4c85-96aa-d596d66762f2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.052148] env[63175]: DEBUG nova.compute.manager [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 963.052488] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.053238] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 963.058984] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652ac832-2ec5-4e62-9902-4ea7147f066e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.062155] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e29f799-2378-4e6b-8f2a-f7d2aa8b8552 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.075678] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93663d8a-3f15-4873-ad16-abbc98c1711e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.082512] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.082512] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efff5c7b-82ac-4598-9fc8-a14e38da2496 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.099203] env[63175]: DEBUG oslo_vmware.api [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 963.099203] env[63175]: value = "task-1248423" [ 963.099203] env[63175]: _type = "Task" [ 963.099203] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.099999] env[63175]: DEBUG nova.compute.provider_tree [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.111117] env[63175]: DEBUG oslo_vmware.api [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248423, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.170036] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Successfully created port: 1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.289607] env[63175]: DEBUG nova.network.neutron [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Updated VIF entry in instance network info cache for port 9f3d24b2-7d86-4548-a817-f8587b820d00. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.289998] env[63175]: DEBUG nova.network.neutron [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating instance_info_cache with network_info: [{"id": "9f3d24b2-7d86-4548-a817-f8587b820d00", "address": "fa:16:3e:ee:45:0c", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f3d24b2-7d", "ovs_interfaceid": "9f3d24b2-7d86-4548-a817-f8587b820d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.326826] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 963.496214] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248422, 'name': CreateVM_Task, 'duration_secs': 0.372276} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.496505] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 572a3821-7436-487a-a053-3819411de57e] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 963.497454] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.497696] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.498092] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.498365] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b40f5d1-8929-4965-bb1e-135306c92450 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.503442] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 963.503442] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5299c17c-e857-3864-9480-24ed78304f2c" [ 963.503442] env[63175]: _type = "Task" [ 963.503442] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.513433] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5299c17c-e857-3864-9480-24ed78304f2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.580277] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.606157] env[63175]: DEBUG nova.scheduler.client.report [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 963.615451] env[63175]: DEBUG oslo_vmware.api [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248423, 'name': PowerOffVM_Task, 'duration_secs': 0.233386} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.615888] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.615888] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.616322] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94c014dd-9bb0-4479-b026-b448ff16b946 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.646608] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.646923] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.647087] env[63175]: DEBUG nova.network.neutron [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.699439] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 963.699669] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 963.699833] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleting the datastore file [datastore2] 089ce349-ac2d-4752-a41a-562b4acd2e73 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.700130] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f68128f-67a2-4bd3-965b-ad855234e729 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.708083] env[63175]: DEBUG oslo_vmware.api [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for the task: (returnval){ [ 963.708083] env[63175]: value = "task-1248425" [ 963.708083] env[63175]: _type = "Task" [ 963.708083] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.793378] env[63175]: DEBUG oslo_concurrency.lockutils [req-0b903458-a69a-4969-988b-bfeee34ec353 req-b2cc9415-71cf-4817-a0b1-9c6df126d467 service nova] Releasing lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.013998] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5299c17c-e857-3864-9480-24ed78304f2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010215} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.014354] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.014605] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.014841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.014994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.015218] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.015495] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2407aa4f-40b2-421e-8193-1ea4465d6a11 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.023555] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.023735] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 964.024474] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ce8739f-640f-4347-82af-0df5323c5ec9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.029901] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 964.029901] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529f6276-3e15-0170-e958-7a1c3673f3dc" [ 964.029901] env[63175]: _type = "Task" [ 964.029901] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.037311] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529f6276-3e15-0170-e958-7a1c3673f3dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.111572] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.113909] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.534s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.115547] env[63175]: INFO nova.compute.claims [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.131084] env[63175]: INFO nova.scheduler.client.report [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Deleted allocations for instance 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f [ 964.218283] env[63175]: DEBUG oslo_vmware.api [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Task: {'id': task-1248425, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153044} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.218484] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.218642] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.218823] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.219014] env[63175]: INFO nova.compute.manager [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Took 1.17 seconds to destroy the instance on the hypervisor. [ 964.219267] env[63175]: DEBUG oslo.service.loopingcall [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.219469] env[63175]: DEBUG nova.compute.manager [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 964.219567] env[63175]: DEBUG nova.network.neutron [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.336845] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 964.362479] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='40de5f611ef314f0c7885cafbf3b942f',container_format='bare',created_at=2024-10-10T14:51:45Z,direct_url=,disk_format='vmdk',id=8b773796-066f-4ad9-8133-ee8b138a0adc,min_disk=1,min_ram=0,name='tempest-test-snap-24114352',owner='83fa8f008a48484195c88ed76ceece0e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T14:51:59Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.362780] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.363032] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.363387] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.363579] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.363758] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.364102] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.364260] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.364460] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.364627] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.364856] env[63175]: DEBUG nova.virt.hardware [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.365816] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c875877a-e99d-4cab-8bdb-cc839f9397f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.374869] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270fbc24-ed45-4eff-abad-96bce69372d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.397712] env[63175]: DEBUG nova.network.neutron [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.457059] env[63175]: DEBUG nova.compute.manager [req-2137eb89-1b42-4b5d-95b4-0e1597f238eb req-b96cfc83-2519-4e80-b1c5-b784a984809c service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Received event network-vif-deleted-23c61658-90f5-4d6c-9db5-24b05bd5e37c {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 964.457176] env[63175]: INFO nova.compute.manager [req-2137eb89-1b42-4b5d-95b4-0e1597f238eb req-b96cfc83-2519-4e80-b1c5-b784a984809c service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Neutron deleted interface 23c61658-90f5-4d6c-9db5-24b05bd5e37c; detaching it from the instance and deleting it from the info cache [ 964.457356] env[63175]: DEBUG nova.network.neutron [req-2137eb89-1b42-4b5d-95b4-0e1597f238eb req-b96cfc83-2519-4e80-b1c5-b784a984809c service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.541463] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529f6276-3e15-0170-e958-7a1c3673f3dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008849} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.542303] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de92f9c2-e66d-4212-b776-ab3b4bc03dc6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.548225] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 964.548225] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf9bfb-d69e-6a85-0246-142221ad99bb" [ 964.548225] env[63175]: _type = "Task" [ 964.548225] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.556070] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf9bfb-d69e-6a85-0246-142221ad99bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.640632] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a43c59e8-3119-464c-9f50-c37ca97e6035 tempest-AttachVolumeShelveTestJSON-879521545 tempest-AttachVolumeShelveTestJSON-879521545-project-member] Lock "5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.398s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.828738] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Successfully updated port: 1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 964.900189] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.936308] env[63175]: DEBUG nova.network.neutron [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.960531] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e146fa-767b-4269-9f7f-39cfd59e4163 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.972190] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa89c72-b181-4285-843e-7f681baf2abf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.001685] env[63175]: DEBUG nova.compute.manager [req-2137eb89-1b42-4b5d-95b4-0e1597f238eb req-b96cfc83-2519-4e80-b1c5-b784a984809c service nova] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Detach interface failed, port_id=23c61658-90f5-4d6c-9db5-24b05bd5e37c, reason: Instance 089ce349-ac2d-4752-a41a-562b4acd2e73 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 965.059040] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bf9bfb-d69e-6a85-0246-142221ad99bb, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.059382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.059680] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 572a3821-7436-487a-a053-3819411de57e/572a3821-7436-487a-a053-3819411de57e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.059959] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fd78aaf-ccc6-493d-9dd9-d6d473c06e0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.067943] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 965.067943] env[63175]: value = "task-1248426" [ 965.067943] env[63175]: _type = "Task" [ 965.067943] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.077792] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.283691] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5facf6eb-848e-46ff-9598-4941e336a198 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.294578] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e507930a-6f29-4bd7-aae0-62f3135173bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.329818] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1538f104-18f9-42bc-8b65-b3cfed6249e8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.332861] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.332967] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.333156] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.340636] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f7b71e-3b47-49ab-94bb-0cd3318756f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.358744] env[63175]: DEBUG nova.compute.provider_tree [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.422949] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8d4056-d8bc-42be-9d34-115463a06c7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.445538] env[63175]: INFO nova.compute.manager [-] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Took 1.23 seconds to deallocate network for instance. [ 965.447525] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28e3f5d-1f1c-48ef-bac7-b50b2a08f06d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.461145] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 83 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.579107] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439097} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.579440] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 572a3821-7436-487a-a053-3819411de57e/572a3821-7436-487a-a053-3819411de57e.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 965.579662] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.579913] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f29af076-142c-4663-aa1c-3434b5515f3a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.586651] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 965.586651] env[63175]: value = "task-1248427" [ 965.586651] env[63175]: _type = "Task" [ 965.586651] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.594380] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.862061] env[63175]: DEBUG nova.scheduler.client.report [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 965.865404] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.955764] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.967816] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.970730] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c8c14a9-b2f4-475b-bae1-f316c55b1f5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.979208] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 965.979208] env[63175]: value = "task-1248428" [ 965.979208] env[63175]: _type = "Task" [ 965.979208] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.988581] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248428, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.053633] env[63175]: DEBUG nova.network.neutron [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Updating instance_info_cache with network_info: [{"id": "1f0cff69-f349-4254-adf9-af30b5209585", "address": "fa:16:3e:13:a9:25", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f0cff69-f3", "ovs_interfaceid": "1f0cff69-f349-4254-adf9-af30b5209585", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.097481] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061154} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.097827] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 966.098928] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31eb64e0-b237-45ff-9dfe-6851eb8ae2a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.121026] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 572a3821-7436-487a-a053-3819411de57e/572a3821-7436-487a-a053-3819411de57e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.121318] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea7ac3c2-721c-4842-9fb2-f3a3ab183e12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.142613] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 966.142613] env[63175]: value = "task-1248429" [ 966.142613] env[63175]: _type = "Task" [ 966.142613] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.151577] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248429, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.368559] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.369135] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 966.372559] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.417s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.372893] env[63175]: DEBUG nova.objects.instance [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lazy-loading 'resources' on Instance uuid 089ce349-ac2d-4752-a41a-562b4acd2e73 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.486179] env[63175]: DEBUG nova.compute.manager [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Received event network-vif-plugged-1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 966.486410] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Acquiring lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.486629] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.486798] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.486972] env[63175]: DEBUG nova.compute.manager [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] No waiting events found dispatching network-vif-plugged-1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 966.487176] env[63175]: WARNING nova.compute.manager [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Received unexpected event network-vif-plugged-1f0cff69-f349-4254-adf9-af30b5209585 for instance with vm_state building and task_state spawning. [ 966.487385] env[63175]: DEBUG nova.compute.manager [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Received event network-changed-1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 966.487552] env[63175]: DEBUG nova.compute.manager [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Refreshing instance network info cache due to event network-changed-1f0cff69-f349-4254-adf9-af30b5209585. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 966.487719] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Acquiring lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.494052] env[63175]: DEBUG oslo_vmware.api [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248428, 'name': PowerOnVM_Task, 'duration_secs': 0.400266} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.494332] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.494565] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d68cd7a-b36a-4f37-8b88-08c29d733e74 tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance '0633c8e9-1c18-4047-ae5d-012d715e03df' progress to 100 {{(pid=63175) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.557090] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.557521] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Instance network_info: |[{"id": "1f0cff69-f349-4254-adf9-af30b5209585", "address": "fa:16:3e:13:a9:25", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f0cff69-f3", "ovs_interfaceid": "1f0cff69-f349-4254-adf9-af30b5209585", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 966.558090] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Acquired lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.558295] env[63175]: DEBUG nova.network.neutron [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Refreshing network info cache for port 1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.559460] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:a9:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '721c6720-3ce0-450e-9951-a894f03acc27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f0cff69-f349-4254-adf9-af30b5209585', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.567350] env[63175]: DEBUG oslo.service.loopingcall [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.570246] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 966.570727] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-996cb962-348b-4046-a947-ab6e4fd834d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.591867] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.591867] env[63175]: value = "task-1248430" [ 966.591867] env[63175]: _type = "Task" [ 966.591867] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.601270] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248430, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.653073] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248429, 'name': ReconfigVM_Task, 'duration_secs': 0.282218} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.653375] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 572a3821-7436-487a-a053-3819411de57e/572a3821-7436-487a-a053-3819411de57e.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.656110] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec7e1124-ca8a-4138-b11e-466b6bc8d969 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.662078] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 966.662078] env[63175]: value = "task-1248431" [ 966.662078] env[63175]: _type = "Task" [ 966.662078] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.674280] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248431, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.849614] env[63175]: DEBUG nova.network.neutron [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Updated VIF entry in instance network info cache for port 1f0cff69-f349-4254-adf9-af30b5209585. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.850035] env[63175]: DEBUG nova.network.neutron [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Updating instance_info_cache with network_info: [{"id": "1f0cff69-f349-4254-adf9-af30b5209585", "address": "fa:16:3e:13:a9:25", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f0cff69-f3", "ovs_interfaceid": "1f0cff69-f349-4254-adf9-af30b5209585", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.876607] env[63175]: DEBUG nova.compute.utils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.877981] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 966.878157] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.926884] env[63175]: DEBUG nova.policy [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.056779] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70750a10-e04e-4a86-a505-22624bd55524 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.071020] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28010f9-40ee-4864-8e64-2d6300ced4cf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.107286] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614360a8-1147-4085-bc06-cf3897ae37b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.117896] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248430, 'name': CreateVM_Task, 'duration_secs': 0.476487} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.119113] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08934c47-15ba-46cf-bfea-56755d0c3d5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.122874] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.123583] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.123756] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.124202] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.124881] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ed8367a-e9c0-4998-802b-e9753bd3b9d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.135429] env[63175]: DEBUG nova.compute.provider_tree [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.137826] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 967.137826] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527b874a-6775-8aef-96dd-5497af81bbe6" [ 967.137826] env[63175]: _type = "Task" [ 967.137826] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.147586] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.147837] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Processing image 8b773796-066f-4ad9-8133-ee8b138a0adc {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.148129] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.148297] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.148483] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.148974] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d52ae4d6-d407-41ac-a329-977c0ba5c0f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.159934] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.160133] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.160836] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31ce9ca9-f534-4436-9851-3e506ba98bdf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.170030] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 967.170030] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52855888-553c-021b-8c4a-4491e9b9dd29" [ 967.170030] env[63175]: _type = "Task" [ 967.170030] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.173572] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248431, 'name': Rename_Task, 'duration_secs': 0.207345} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.176361] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.176475] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5df636bd-d684-4185-9c54-748242c05b12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.187778] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52855888-553c-021b-8c4a-4491e9b9dd29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.187778] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 967.187778] env[63175]: value = "task-1248432" [ 967.187778] env[63175]: _type = "Task" [ 967.187778] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.194374] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.314317] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Successfully created port: aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.352927] env[63175]: DEBUG oslo_concurrency.lockutils [req-9a2eefd7-d24f-4128-a84b-4fdacd644233 req-83e64518-ca19-47fc-982f-3309babfe45c service nova] Releasing lock "refresh_cache-63be6b3c-2535-4c26-9a15-9c2445b049e8" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.387300] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 967.640026] env[63175]: DEBUG nova.scheduler.client.report [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 967.683253] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Preparing fetch location {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 967.683547] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Fetch image to [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481/OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481.vmdk {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 967.683739] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Downloading stream optimized image 8b773796-066f-4ad9-8133-ee8b138a0adc to [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481/OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481.vmdk on the data store datastore2 as vApp {{(pid=63175) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 967.683919] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Downloading image file data 8b773796-066f-4ad9-8133-ee8b138a0adc to the ESX as VM named 'OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481' {{(pid=63175) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 967.698367] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248432, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.769746] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 967.769746] env[63175]: value = "resgroup-9" [ 967.769746] env[63175]: _type = "ResourcePool" [ 967.769746] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 967.770228] env[63175]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1acda8fa-de91-406e-b9c0-62f96c791be7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.792307] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease: (returnval){ [ 967.792307] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527efbf9-b13d-882b-1766-a3665af80bb4" [ 967.792307] env[63175]: _type = "HttpNfcLease" [ 967.792307] env[63175]: } obtained for vApp import into resource pool (val){ [ 967.792307] env[63175]: value = "resgroup-9" [ 967.792307] env[63175]: _type = "ResourcePool" [ 967.792307] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 967.792566] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the lease: (returnval){ [ 967.792566] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527efbf9-b13d-882b-1766-a3665af80bb4" [ 967.792566] env[63175]: _type = "HttpNfcLease" [ 967.792566] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 967.801728] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 967.801728] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527efbf9-b13d-882b-1766-a3665af80bb4" [ 967.801728] env[63175]: _type = "HttpNfcLease" [ 967.801728] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.019030] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.019030] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.145120] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.167765] env[63175]: INFO nova.scheduler.client.report [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Deleted allocations for instance 089ce349-ac2d-4752-a41a-562b4acd2e73 [ 968.197900] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248432, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.303360] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.303360] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527efbf9-b13d-882b-1766-a3665af80bb4" [ 968.303360] env[63175]: _type = "HttpNfcLease" [ 968.303360] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 968.303794] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 968.303794] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527efbf9-b13d-882b-1766-a3665af80bb4" [ 968.303794] env[63175]: _type = "HttpNfcLease" [ 968.303794] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 968.304864] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd6d37a-b5b3-4054-82ac-0cfe26a3bf00 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.312385] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 968.312577] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk. {{(pid=63175) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 968.379112] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e53ec805-7c05-4c4f-93b8-e7a712737760 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.393722] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 968.417448] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.417741] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.417923] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.418145] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.418307] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.418480] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.418721] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.418907] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.419116] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.419325] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.419536] env[63175]: DEBUG nova.virt.hardware [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.420445] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7ea1d2-35dc-4c28-ba90-19cffd500aac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.429400] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ee8486-fadf-4984-b1c4-a6b49c06f31a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.484372] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.484683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.484889] env[63175]: DEBUG nova.compute.manager [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Going to confirm migration 3 {{(pid=63175) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 968.526710] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 968.526903] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 968.680119] env[63175]: DEBUG oslo_concurrency.lockutils [None req-688a9e27-a553-4350-8710-e82c2650a50a tempest-ServerDiskConfigTestJSON-142080242 tempest-ServerDiskConfigTestJSON-142080242-project-member] Lock "089ce349-ac2d-4752-a41a-562b4acd2e73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.152s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.703750] env[63175]: DEBUG oslo_vmware.api [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248432, 'name': PowerOnVM_Task, 'duration_secs': 1.461494} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.706115] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.706353] env[63175]: INFO nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Took 8.41 seconds to spawn the instance on the hypervisor. [ 968.706542] env[63175]: DEBUG nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 968.707617] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c5f83b-ec4c-4cc5-bbc3-df773891bcce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.804657] env[63175]: DEBUG nova.compute.manager [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-vif-plugged-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 968.804938] env[63175]: DEBUG oslo_concurrency.lockutils [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.805396] env[63175]: DEBUG oslo_concurrency.lockutils [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.805605] env[63175]: DEBUG oslo_concurrency.lockutils [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.805837] env[63175]: DEBUG nova.compute.manager [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] No waiting events found dispatching network-vif-plugged-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 968.806035] env[63175]: WARNING nova.compute.manager [req-d11a1e52-46e3-45d4-93fa-18cd1411c7ed req-08f94529-c0fd-44be-85c9-02e9e6d14d92 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received unexpected event network-vif-plugged-aab7c934-a283-4031-bf42-478ef430f7ef for instance with vm_state building and task_state spawning. [ 968.905547] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Successfully updated port: aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.058992] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.059116] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquired lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.059365] env[63175]: DEBUG nova.network.neutron [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.060034] env[63175]: DEBUG nova.objects.instance [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'info_cache' on Instance uuid 0633c8e9-1c18-4047-ae5d-012d715e03df {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.156876] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Completed reading data from the image iterator. {{(pid=63175) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 969.157119] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 969.158041] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d6884c-2628-4ce5-9b5a-ec6b08d4edcd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.165638] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 969.165878] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 969.166110] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-26ba9a4f-7bb5-4d88-ba9a-b7f4a70285f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.229296] env[63175]: INFO nova.compute.manager [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Took 13.24 seconds to build instance. [ 969.439965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.439965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.439965] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.446400] env[63175]: DEBUG oslo_vmware.rw_handles [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52acd30a-ae6c-18ba-1ced-3963964772d2/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 969.446400] env[63175]: INFO nova.virt.vmwareapi.images [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Downloaded image file data 8b773796-066f-4ad9-8133-ee8b138a0adc [ 969.446400] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa633cc-69a6-4a44-83e8-458fae83a5b1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.460203] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08f18e53-696e-4dd7-9dd8-010a49af6421 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.492127] env[63175]: INFO nova.virt.vmwareapi.images [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] The imported VM was unregistered [ 969.495150] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Caching image {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 969.495150] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.495332] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-410f3f17-a804-49f9-a838-d770511b1e45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.527510] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.527729] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481/OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481.vmdk to [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk. {{(pid=63175) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 969.528013] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d49879fc-eea6-4396-81ed-ef8f27e0744b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.536702] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 969.536702] env[63175]: value = "task-1248435" [ 969.536702] env[63175]: _type = "Task" [ 969.536702] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.545155] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.730853] env[63175]: DEBUG oslo_concurrency.lockutils [None req-aeb2c37d-8584-454f-8594-6f6dd4d32415 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.749s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.970474] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.052234] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.267085] env[63175]: DEBUG nova.network.neutron [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.388806] env[63175]: DEBUG nova.network.neutron [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [{"id": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "address": "fa:16:3e:b3:43:aa", "network": {"id": "d67ca385-0d8d-4a4d-bed0-8395d4c52d0d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-127460578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "71ca546385844c0d803034ef9e853377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4621f750-c3", "ovs_interfaceid": "4621f750-c3ff-4289-bf4f-9ad952a25f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.555700] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.774151] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.774978] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Instance network_info: |[{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 970.775210] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:bb:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aab7c934-a283-4031-bf42-478ef430f7ef', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.785535] env[63175]: DEBUG oslo.service.loopingcall [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.785731] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.786016] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-863adde2-10cd-404b-9d99-d5f455f30919 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.819544] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.819544] env[63175]: value = "task-1248436" [ 970.819544] env[63175]: _type = "Task" [ 970.819544] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.834099] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248436, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.893656] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Releasing lock "refresh_cache-0633c8e9-1c18-4047-ae5d-012d715e03df" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.894046] env[63175]: DEBUG nova.objects.instance [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'migration_context' on Instance uuid 0633c8e9-1c18-4047-ae5d-012d715e03df {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.938228] env[63175]: DEBUG nova.compute.manager [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 970.938543] env[63175]: DEBUG nova.compute.manager [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 970.938793] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.938961] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.939816] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.057871] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.335069] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248436, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.397786] env[63175]: DEBUG nova.objects.base [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Object Instance<0633c8e9-1c18-4047-ae5d-012d715e03df> lazy-loaded attributes: info_cache,migration_context {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 971.399091] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a1bf3f-b5cd-41ad-a874-d08ec3694442 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.424802] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279b022e-245f-4a02-bfe7-f8dc24a09c7d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.434496] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 971.434496] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52341be8-617e-492d-597d-cb8ebd5419e2" [ 971.434496] env[63175]: _type = "Task" [ 971.434496] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.452802] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52341be8-617e-492d-597d-cb8ebd5419e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.556843] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.604353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.604660] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquired lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.604874] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Forcefully refreshing network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 971.724105] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.724715] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.837107] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248436, 'name': CreateVM_Task, 'duration_secs': 0.526616} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.837107] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.837567] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.837709] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.838036] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.838366] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d276e1a-374d-46d0-aac5-3cfce5d53874 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.848321] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 971.848321] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523118f0-cdea-862d-5c87-2ffdb4af18ef" [ 971.848321] env[63175]: _type = "Task" [ 971.848321] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.862010] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523118f0-cdea-862d-5c87-2ffdb4af18ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.950414] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52341be8-617e-492d-597d-cb8ebd5419e2, 'name': SearchDatastore_Task, 'duration_secs': 0.049394} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.950741] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.950983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.057599] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.227959] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.228488] env[63175]: DEBUG nova.compute.manager [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Received event network-changed-9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 972.228948] env[63175]: DEBUG nova.compute.manager [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Refreshing instance network info cache due to event network-changed-9f3d24b2-7d86-4548-a817-f8587b820d00. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 972.228948] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Acquiring lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.229086] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Acquired lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.229299] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Refreshing network info cache for port 9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.359932] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523118f0-cdea-862d-5c87-2ffdb4af18ef, 'name': SearchDatastore_Task, 'duration_secs': 0.088336} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.360230] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.360477] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.360713] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.360866] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.361067] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.361344] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31ff090b-7b81-442f-8001-5fce72217051 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.382018] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.382018] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.382537] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beb2c802-fd3a-46a4-a797-f0bb1bd3bac0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.388482] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 972.388482] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5206cc72-42af-c72a-04da-40922432cde3" [ 972.388482] env[63175]: _type = "Task" [ 972.388482] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.396310] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5206cc72-42af-c72a-04da-40922432cde3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.551716] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.587876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ab8abe-b166-4819-87be-9fb756f5a673 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.595826] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619567c7-f1bd-4cba-a430-290acd14500c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.629749] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88827e04-17d4-40df-8d3c-427bb73d3561 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.637770] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17921e3a-24f6-49eb-9109-7d341c9c371b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.654854] env[63175]: DEBUG nova.compute.provider_tree [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.899188] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [{"id": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "address": "fa:16:3e:45:57:5f", "network": {"id": "0d4031c6-1418-4b35-9dd1-345b6e71f68b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-314297824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822ae9f380c84b3c914770f1232975b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0467beaa-08c6-44d6-b8a2-e9c609c21ff4", "external-id": "nsx-vlan-transportzone-540", "segmentation_id": 540, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25225abe-2d", "ovs_interfaceid": "25225abe-2dcf-4815-85da-5fbf5ea4f2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.905389] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5206cc72-42af-c72a-04da-40922432cde3, 'name': SearchDatastore_Task, 'duration_secs': 0.082524} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.906720] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e34f7bdf-0f01-4d5b-8608-b8deed330157 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.915650] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 972.915650] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dca581-9820-feac-e7ca-2c3703f359fe" [ 972.915650] env[63175]: _type = "Task" [ 972.915650] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.926424] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dca581-9820-feac-e7ca-2c3703f359fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.030765] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Updated VIF entry in instance network info cache for port 9f3d24b2-7d86-4548-a817-f8587b820d00. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 973.031191] env[63175]: DEBUG nova.network.neutron [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating instance_info_cache with network_info: [{"id": "9f3d24b2-7d86-4548-a817-f8587b820d00", "address": "fa:16:3e:ee:45:0c", "network": {"id": "7135995a-866b-453b-97ba-987af3e8b015", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1158677553-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8826d0f0384b56aa862756787ad017", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f3d24b2-7d", "ovs_interfaceid": "9f3d24b2-7d86-4548-a817-f8587b820d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.054907] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248435, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.399551} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.055629] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481/OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481.vmdk to [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk. [ 973.056139] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Cleaning up location [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 973.056139] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_1a1b916f-5168-4db0-8904-9532a6b29481 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.056605] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0eb51b12-3db4-4b5f-96cb-3fcc25a0aa49 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.065617] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 973.065617] env[63175]: value = "task-1248437" [ 973.065617] env[63175]: _type = "Task" [ 973.065617] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.075583] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.157771] env[63175]: DEBUG nova.scheduler.client.report [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 973.407241] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Releasing lock "refresh_cache-a460926e-9637-40aa-bb30-e3890a441e03" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.407632] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updated the network info_cache for instance {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 973.407898] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.408648] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.408867] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.409068] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.409205] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.409359] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.409492] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 973.409637] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.427045] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dca581-9820-feac-e7ca-2c3703f359fe, 'name': SearchDatastore_Task, 'duration_secs': 0.022133} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.427316] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.427574] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 48448a68-6b7e-488b-a7f2-9f910a2f6765/48448a68-6b7e-488b-a7f2-9f910a2f6765.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.428079] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5ecc75b-ff48-4aef-a083-86b41c12ea15 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.435897] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 973.435897] env[63175]: value = "task-1248438" [ 973.435897] env[63175]: _type = "Task" [ 973.435897] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.444578] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248438, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.534490] env[63175]: DEBUG oslo_concurrency.lockutils [req-3868f57b-0342-48ed-8a32-20a7f53d19cb req-3039297c-4ac3-46d0-973a-7d67e641b150 service nova] Releasing lock "refresh_cache-572a3821-7436-487a-a053-3819411de57e" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.577490] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18997} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.577886] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.578164] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.578532] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk to [datastore2] 63be6b3c-2535-4c26-9a15-9c2445b049e8/63be6b3c-2535-4c26-9a15-9c2445b049e8.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.578877] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76d2987f-2104-48e9-8f48-0108cd5e6e46 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.587486] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 973.587486] env[63175]: value = "task-1248439" [ 973.587486] env[63175]: _type = "Task" [ 973.587486] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.597589] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.912400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.946986] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248438, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.101312] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.168940] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.218s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.177026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.261s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.177026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.177026] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 974.177026] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dba4a0d-e691-4d95-91b8-6280089913ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.185876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36e5b70-7f31-467d-9c65-77fc4a5bf846 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.210700] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85d522d-63cb-46a8-8835-0dcb72aa12b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.223706] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4715f4c-f618-4047-b5c8-a57ace2b74d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.262779] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180118MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 974.265959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.266507] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.003s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.453280] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248438, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.601292] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.724847] env[63175]: INFO nova.scheduler.client.report [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocation for migration dbd8eafe-cfcc-4812-b66e-45973401f12f [ 974.953870] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248438, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.113794} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.954267] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 48448a68-6b7e-488b-a7f2-9f910a2f6765/48448a68-6b7e-488b-a7f2-9f910a2f6765.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.954543] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.954791] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68330e20-54c9-492c-84b7-863c51a7424c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.967874] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 974.967874] env[63175]: value = "task-1248440" [ 974.967874] env[63175]: _type = "Task" [ 974.967874] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.982917] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248440, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.102997] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.133053] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.133314] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.230964] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.746s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.298291] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a460926e-9637-40aa-bb30-e3890a441e03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.298501] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.298569] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0645fffe-9910-43af-af40-126592aefaff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.298686] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.298797] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.298906] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 0633c8e9-1c18-4047-ae5d-012d715e03df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.299028] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 572a3821-7436-487a-a053-3819411de57e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.299144] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 63be6b3c-2535-4c26-9a15-9c2445b049e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.299252] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 48448a68-6b7e-488b-a7f2-9f910a2f6765 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 975.483483] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248440, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077244} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.483807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.484755] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b075fc78-5c0d-4495-bfbc-2976209bad7a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.512472] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 48448a68-6b7e-488b-a7f2-9f910a2f6765/48448a68-6b7e-488b-a7f2-9f910a2f6765.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.512858] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c66b083f-626a-4052-a9a0-30036d669a05 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.538412] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 975.538412] env[63175]: value = "task-1248441" [ 975.538412] env[63175]: _type = "Task" [ 975.538412] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.551714] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248441, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.602447] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.636217] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 975.802078] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 975.802542] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 975.802681] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 975.972507] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0b776a-db70-451d-8610-a42dab5da5a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.985428] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a60055-491e-467c-a24a-de3c0ee682cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.027608] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c99969-c6a4-47e9-b23e-39896af6614f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.047313] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307260aa-31ba-494d-aaf6-d2b8ff30ce5d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.059080] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248441, 'name': ReconfigVM_Task, 'duration_secs': 0.416664} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.069318] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 48448a68-6b7e-488b-a7f2-9f910a2f6765/48448a68-6b7e-488b-a7f2-9f910a2f6765.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.070682] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.071941] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9415d6ff-93d3-4622-8be3-80453b6dba60 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.084734] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 976.084734] env[63175]: value = "task-1248442" [ 976.084734] env[63175]: _type = "Task" [ 976.084734] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.099271] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248442, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.102859] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.158937] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.534720] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.534967] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.535153] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.535427] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.535555] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.537863] env[63175]: INFO nova.compute.manager [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Terminating instance [ 976.576203] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 976.597786] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248442, 'name': Rename_Task, 'duration_secs': 0.174716} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.598518] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.598981] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee235f27-3a8c-412f-af81-482937bd5d63 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.603915] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248439, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.886528} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.604526] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8b773796-066f-4ad9-8133-ee8b138a0adc/8b773796-066f-4ad9-8133-ee8b138a0adc.vmdk to [datastore2] 63be6b3c-2535-4c26-9a15-9c2445b049e8/63be6b3c-2535-4c26-9a15-9c2445b049e8.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.605370] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2de9fb-1d97-4393-a0cf-b785ec46d3a3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.609826] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 976.609826] env[63175]: value = "task-1248443" [ 976.609826] env[63175]: _type = "Task" [ 976.609826] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.631378] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 63be6b3c-2535-4c26-9a15-9c2445b049e8/63be6b3c-2535-4c26-9a15-9c2445b049e8.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.632512] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c59f6444-edcc-44ca-8caa-0a8d9b0900f2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.653654] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248443, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.660525] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 976.660525] env[63175]: value = "task-1248444" [ 976.660525] env[63175]: _type = "Task" [ 976.660525] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.669671] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248444, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.041808] env[63175]: DEBUG nova.compute.manager [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 977.042179] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 977.043238] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da73150c-f7fb-425f-970a-9ec4d2afae59 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.051502] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.051762] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b17e6f50-d038-431f-8a64-fe18ce8f3534 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.060139] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 977.060139] env[63175]: value = "task-1248445" [ 977.060139] env[63175]: _type = "Task" [ 977.060139] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.069240] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.080830] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 977.081159] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.815s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.081588] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.923s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.084011] env[63175]: INFO nova.compute.claims [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.120924] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248443, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.171275] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.571309] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248445, 'name': PowerOffVM_Task, 'duration_secs': 0.501219} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.571505] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.572450] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.572450] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ec91c95-435e-4640-8fea-e0f038704fa8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.625260] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248443, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.673953] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.680590] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.681064] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.681407] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleting the datastore file [datastore2] 0633c8e9-1c18-4047-ae5d-012d715e03df {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.681754] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-723fd45d-b161-425d-a80b-78ea447fe4dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.691212] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for the task: (returnval){ [ 977.691212] env[63175]: value = "task-1248447" [ 977.691212] env[63175]: _type = "Task" [ 977.691212] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.700846] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.122311] env[63175]: DEBUG oslo_vmware.api [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248443, 'name': PowerOnVM_Task, 'duration_secs': 1.079708} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.125242] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.125524] env[63175]: INFO nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Took 9.73 seconds to spawn the instance on the hypervisor. [ 978.125733] env[63175]: DEBUG nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 978.126785] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c26ff00-a878-421a-9bd9-5fbddfe9770e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.175792] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248444, 'name': ReconfigVM_Task, 'duration_secs': 1.449296} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.176124] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 63be6b3c-2535-4c26-9a15-9c2445b049e8/63be6b3c-2535-4c26-9a15-9c2445b049e8.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.177062] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b16bf7f-111e-441b-9a9e-fd70ab2f5784 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.190237] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 978.190237] env[63175]: value = "task-1248448" [ 978.190237] env[63175]: _type = "Task" [ 978.190237] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.204663] env[63175]: DEBUG oslo_vmware.api [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Task: {'id': task-1248447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430867} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.207911] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.208144] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.208339] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.208526] env[63175]: INFO nova.compute.manager [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Took 1.17 seconds to destroy the instance on the hypervisor. [ 978.208819] env[63175]: DEBUG oslo.service.loopingcall [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.209049] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248448, 'name': Rename_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.211477] env[63175]: DEBUG nova.compute.manager [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 978.211579] env[63175]: DEBUG nova.network.neutron [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.236438] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e9a6c0-47a3-4443-890b-30b08be3d749 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.247051] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7a30c9-a16d-4bb5-a68f-1c17b6bd6912 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.294304] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56cc48d7-2636-40f9-a52e-d689cf6e39a5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.304609] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c48cdc-f3b9-4558-af9d-22e28da51580 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.319524] env[63175]: DEBUG nova.compute.provider_tree [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.534969] env[63175]: DEBUG nova.compute.manager [req-7f3fc517-7a2f-4b65-ba21-05dbccbb8ed8 req-51121c5d-22ac-4bbf-9f41-e0b5730d634a service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Received event network-vif-deleted-4621f750-c3ff-4289-bf4f-9ad952a25f05 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 978.534969] env[63175]: INFO nova.compute.manager [req-7f3fc517-7a2f-4b65-ba21-05dbccbb8ed8 req-51121c5d-22ac-4bbf-9f41-e0b5730d634a service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Neutron deleted interface 4621f750-c3ff-4289-bf4f-9ad952a25f05; detaching it from the instance and deleting it from the info cache [ 978.535162] env[63175]: DEBUG nova.network.neutron [req-7f3fc517-7a2f-4b65-ba21-05dbccbb8ed8 req-51121c5d-22ac-4bbf-9f41-e0b5730d634a service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.648660] env[63175]: INFO nova.compute.manager [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Took 15.08 seconds to build instance. [ 978.700743] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248448, 'name': Rename_Task, 'duration_secs': 0.149911} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.701044] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.701299] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fb39bc3-edad-4a78-b54a-705221e3c809 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.725160] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 978.725160] env[63175]: value = "task-1248449" [ 978.725160] env[63175]: _type = "Task" [ 978.725160] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.734949] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.823363] env[63175]: DEBUG nova.scheduler.client.report [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 979.014613] env[63175]: DEBUG nova.network.neutron [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.037466] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-347f360e-11aa-4ecf-a219-3804fb281ef8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.047767] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56db057d-ffb1-4bd9-9ce9-c396fbf8cba9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.077946] env[63175]: DEBUG nova.compute.manager [req-7f3fc517-7a2f-4b65-ba21-05dbccbb8ed8 req-51121c5d-22ac-4bbf-9f41-e0b5730d634a service nova] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Detach interface failed, port_id=4621f750-c3ff-4289-bf4f-9ad952a25f05, reason: Instance 0633c8e9-1c18-4047-ae5d-012d715e03df could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 979.150847] env[63175]: DEBUG oslo_concurrency.lockutils [None req-88a22d08-0f54-4ab9-8275-d51cbf95c015 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.609s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.234788] env[63175]: DEBUG oslo_vmware.api [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248449, 'name': PowerOnVM_Task, 'duration_secs': 0.499922} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.235063] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.235236] env[63175]: INFO nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Took 14.90 seconds to spawn the instance on the hypervisor. [ 979.235445] env[63175]: DEBUG nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 979.236208] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681cd05c-00bb-44bd-93db-a295461e7a7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.328050] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.246s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.328630] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 979.517643] env[63175]: INFO nova.compute.manager [-] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Took 1.31 seconds to deallocate network for instance. [ 979.742707] env[63175]: DEBUG nova.compute.manager [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 979.742995] env[63175]: DEBUG nova.compute.manager [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 979.743141] env[63175]: DEBUG oslo_concurrency.lockutils [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.743296] env[63175]: DEBUG oslo_concurrency.lockutils [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.743466] env[63175]: DEBUG nova.network.neutron [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.754189] env[63175]: INFO nova.compute.manager [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Took 22.29 seconds to build instance. [ 979.835726] env[63175]: DEBUG nova.compute.utils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.837727] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 979.837727] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.883615] env[63175]: DEBUG nova.policy [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '38193549cdf44b13abe7a3fd6b9691e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99d35bb46fce45ed89837f84c54c49c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.024368] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.024674] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.024900] env[63175]: DEBUG nova.objects.instance [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lazy-loading 'resources' on Instance uuid 0633c8e9-1c18-4047-ae5d-012d715e03df {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.111230] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.164777] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Successfully created port: 34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.256356] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e8623ec0-392f-4cf9-b80f-edbe80786317 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.809s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.256711] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.146s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.256831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.257022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.257197] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.259382] env[63175]: INFO nova.compute.manager [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Terminating instance [ 980.340330] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 980.609010] env[63175]: DEBUG nova.network.neutron [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.609398] env[63175]: DEBUG nova.network.neutron [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.630142] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.630378] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.667766] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.668023] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.680983] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8173ca7-cf31-4173-9307-fc311e7267ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.688832] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845ddfca-d4ed-4842-80ad-e93fbde8dd96 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.721462] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1b4f56-6677-4383-a440-9c12a83c1b47 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.729620] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b9b3b6-d114-4acb-8363-b90789863992 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.744693] env[63175]: DEBUG nova.compute.provider_tree [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.764072] env[63175]: DEBUG nova.compute.manager [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 980.764072] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.764626] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea737174-480c-4b7a-8d3a-b5ae23e7400c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.772933] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.773197] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c15aebc-229c-4255-8118-0a87b397d22a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.780327] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 980.780327] env[63175]: value = "task-1248450" [ 980.780327] env[63175]: _type = "Task" [ 980.780327] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.789245] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.113834] env[63175]: DEBUG oslo_concurrency.lockutils [req-59862cf1-696a-4c34-ad92-24c6ede73439 req-d6a1ae14-e939-4322-8ddf-0aa521586ea2 service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.132638] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 981.173027] env[63175]: INFO nova.compute.manager [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Detaching volume 345211ba-3866-4941-b49b-db275c459229 [ 981.207533] env[63175]: INFO nova.virt.block_device [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Attempting to driver detach volume 345211ba-3866-4941-b49b-db275c459229 from mountpoint /dev/sdb [ 981.207788] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 981.207985] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269071', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'name': 'volume-345211ba-3866-4941-b49b-db275c459229', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0645fffe-9910-43af-af40-126592aefaff', 'attached_at': '', 'detached_at': '', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'serial': '345211ba-3866-4941-b49b-db275c459229'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 981.209025] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c05849-6cd5-4d21-b3cf-41cf8f0388b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.230485] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb41895-adab-493c-8baf-8ca76702920b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.238417] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2dc1e3-7efc-48ee-a9bb-9f3c620f4bd0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.260395] env[63175]: DEBUG nova.scheduler.client.report [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 981.264491] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d83ad1-5feb-47f2-a9fb-34a6647b7c46 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.279429] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] The volume has not been displaced from its original location: [datastore2] volume-345211ba-3866-4941-b49b-db275c459229/volume-345211ba-3866-4941-b49b-db275c459229.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 981.284584] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 981.285686] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c28b63f-5eac-4720-8e10-5bd19e4005fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.306285] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248450, 'name': PowerOffVM_Task, 'duration_secs': 0.204732} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.307515] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.307729] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.308093] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 981.308093] env[63175]: value = "task-1248451" [ 981.308093] env[63175]: _type = "Task" [ 981.308093] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.308368] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5630d416-2b34-4a0b-8882-72902fed449c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.320839] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248451, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.349828] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 981.376831] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.377115] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.377287] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.377483] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.377673] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.377830] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.378054] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.378227] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.378403] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.378633] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.378819] env[63175]: DEBUG nova.virt.hardware [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.379805] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdb2e78-fb97-4848-a4ef-85ec586197d6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.383802] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.384014] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.384218] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore2] 63be6b3c-2535-4c26-9a15-9c2445b049e8 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.384962] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17f230b9-f4d5-4a37-ab95-85810e5ac04d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.390318] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743186ec-ccdd-409a-80f2-aa5cb93d3e38 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.395176] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 981.395176] env[63175]: value = "task-1248453" [ 981.395176] env[63175]: _type = "Task" [ 981.395176] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.410994] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248453, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.659968] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.739031] env[63175]: DEBUG nova.compute.manager [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Received event network-vif-plugged-34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 981.739280] env[63175]: DEBUG oslo_concurrency.lockutils [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] Acquiring lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.739531] env[63175]: DEBUG oslo_concurrency.lockutils [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.739661] env[63175]: DEBUG oslo_concurrency.lockutils [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.739835] env[63175]: DEBUG nova.compute.manager [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] No waiting events found dispatching network-vif-plugged-34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 981.740010] env[63175]: WARNING nova.compute.manager [req-de1697ae-5193-469f-b0aa-242ae0c8e8e3 req-874a3fcf-320e-4700-a406-b48518d6b490 service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Received unexpected event network-vif-plugged-34325eb5-e47f-446d-925c-bc3a7e76a2d3 for instance with vm_state building and task_state spawning. [ 981.765547] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.768820] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.108s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.770429] env[63175]: INFO nova.compute.claims [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.789547] env[63175]: INFO nova.scheduler.client.report [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Deleted allocations for instance 0633c8e9-1c18-4047-ae5d-012d715e03df [ 981.824516] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248451, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.832512] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Successfully updated port: 34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.910497] env[63175]: DEBUG oslo_vmware.api [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248453, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137762} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.910923] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.911750] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.912118] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.912598] env[63175]: INFO nova.compute.manager [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 981.912941] env[63175]: DEBUG oslo.service.loopingcall [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.913669] env[63175]: DEBUG nova.compute.manager [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 981.913932] env[63175]: DEBUG nova.network.neutron [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.297462] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c14a891c-7135-4697-8877-80707a94a89e tempest-DeleteServersTestJSON-2112326937 tempest-DeleteServersTestJSON-2112326937-project-member] Lock "0633c8e9-1c18-4047-ae5d-012d715e03df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.762s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.322572] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248451, 'name': ReconfigVM_Task, 'duration_secs': 0.520554} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.322870] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 982.327487] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-894bfe1b-ce8a-4b9d-bf86-26a32b70e481 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.337581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.337715] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquired lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.337868] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.346390] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 982.346390] env[63175]: value = "task-1248454" [ 982.346390] env[63175]: _type = "Task" [ 982.346390] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.354968] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.654082] env[63175]: DEBUG nova.network.neutron [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.858181] env[63175]: DEBUG oslo_vmware.api [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248454, 'name': ReconfigVM_Task, 'duration_secs': 0.237761} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.858511] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269071', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'name': 'volume-345211ba-3866-4941-b49b-db275c459229', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0645fffe-9910-43af-af40-126592aefaff', 'attached_at': '', 'detached_at': '', 'volume_id': '345211ba-3866-4941-b49b-db275c459229', 'serial': '345211ba-3866-4941-b49b-db275c459229'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 982.869742] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.918248] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a28d0f-993d-4dbb-a1af-98581d4497ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.927294] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f7c6bf-d4ab-44e6-8d5d-a446db535e5a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.962515] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dd4af8-eeee-451c-a630-8e1a03666c5f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.970422] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014a4689-d037-4cfa-a926-073e9125c85b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.984590] env[63175]: DEBUG nova.compute.provider_tree [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.044843] env[63175]: DEBUG nova.network.neutron [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Updating instance_info_cache with network_info: [{"id": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "address": "fa:16:3e:38:58:bc", "network": {"id": "d4c19273-0dae-4213-aec1-fd37302f18ea", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-479027390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99d35bb46fce45ed89837f84c54c49c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34325eb5-e4", "ovs_interfaceid": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.158641] env[63175]: INFO nova.compute.manager [-] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Took 1.24 seconds to deallocate network for instance. [ 983.400344] env[63175]: DEBUG nova.objects.instance [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'flavor' on Instance uuid 0645fffe-9910-43af-af40-126592aefaff {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.487510] env[63175]: DEBUG nova.scheduler.client.report [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 983.547559] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Releasing lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.547891] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Instance network_info: |[{"id": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "address": "fa:16:3e:38:58:bc", "network": {"id": "d4c19273-0dae-4213-aec1-fd37302f18ea", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-479027390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99d35bb46fce45ed89837f84c54c49c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34325eb5-e4", "ovs_interfaceid": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 983.548359] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:58:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34325eb5-e47f-446d-925c-bc3a7e76a2d3', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 983.556109] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Creating folder: Project (99d35bb46fce45ed89837f84c54c49c9). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 983.556816] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3e8864b-7be1-47b8-b64a-4d17d38492e9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.569272] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Created folder: Project (99d35bb46fce45ed89837f84c54c49c9) in parent group-v268956. [ 983.569482] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Creating folder: Instances. Parent ref: group-v269082. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 983.569743] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f08b0d6-8cf8-4ed3-9ffa-291df8196149 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.579310] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Created folder: Instances in parent group-v269082. [ 983.579550] env[63175]: DEBUG oslo.service.loopingcall [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.579754] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 983.579966] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1188c020-e4d2-48e4-96fe-b7b852720ad6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.599990] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 983.599990] env[63175]: value = "task-1248458" [ 983.599990] env[63175]: _type = "Task" [ 983.599990] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.608200] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248458, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.664863] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.770039] env[63175]: DEBUG nova.compute.manager [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Received event network-changed-34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 983.770217] env[63175]: DEBUG nova.compute.manager [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Refreshing instance network info cache due to event network-changed-34325eb5-e47f-446d-925c-bc3a7e76a2d3. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 983.770450] env[63175]: DEBUG oslo_concurrency.lockutils [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] Acquiring lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.770606] env[63175]: DEBUG oslo_concurrency.lockutils [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] Acquired lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.770774] env[63175]: DEBUG nova.network.neutron [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Refreshing network info cache for port 34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 983.992771] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.993317] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 983.996016] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.331s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.996246] env[63175]: DEBUG nova.objects.instance [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'resources' on Instance uuid 63be6b3c-2535-4c26-9a15-9c2445b049e8 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.111210] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248458, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.407345] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8b3e6fe1-a250-4de6-856e-8f272c1cb76c tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.739s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.461610] env[63175]: DEBUG nova.network.neutron [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Updated VIF entry in instance network info cache for port 34325eb5-e47f-446d-925c-bc3a7e76a2d3. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.461980] env[63175]: DEBUG nova.network.neutron [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Updating instance_info_cache with network_info: [{"id": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "address": "fa:16:3e:38:58:bc", "network": {"id": "d4c19273-0dae-4213-aec1-fd37302f18ea", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-479027390-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99d35bb46fce45ed89837f84c54c49c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34325eb5-e4", "ovs_interfaceid": "34325eb5-e47f-446d-925c-bc3a7e76a2d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.499593] env[63175]: DEBUG nova.compute.utils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.504258] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 984.504374] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.574252] env[63175]: DEBUG nova.policy [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.611570] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248458, 'name': CreateVM_Task, 'duration_secs': 0.532075} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.613772] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 984.614985] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.614985] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.618460] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 984.618460] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e8b472-a5e7-4985-be95-d03244779b07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.619711] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 984.619711] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5222e1a2-782e-230c-0cab-be31b7401e24" [ 984.619711] env[63175]: _type = "Task" [ 984.619711] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.629322] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5222e1a2-782e-230c-0cab-be31b7401e24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.635695] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41b2fe1-615b-4943-8513-ff69b3756a96 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.643276] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbb1fec-ffcd-4c0c-bf43-533c1e528a32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.674051] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee355439-a788-495c-96ee-a5664af34192 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.681286] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e0b2b1-6efb-4e45-a0ed-94ed2152c00a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.694138] env[63175]: DEBUG nova.compute.provider_tree [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.897898] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Successfully created port: 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.967547] env[63175]: DEBUG oslo_concurrency.lockutils [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] Releasing lock "refresh_cache-9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.968089] env[63175]: DEBUG nova.compute.manager [req-b126d964-9a5d-4abf-ae69-700df2dc6ddb req-2bc9de4e-dd0a-4f3c-aff8-a46e1e133d5d service nova] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Received event network-vif-deleted-1f0cff69-f349-4254-adf9-af30b5209585 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 984.968542] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.969181] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.969424] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "0645fffe-9910-43af-af40-126592aefaff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.969751] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.970055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.972605] env[63175]: INFO nova.compute.manager [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Terminating instance [ 985.003886] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 985.132145] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5222e1a2-782e-230c-0cab-be31b7401e24, 'name': SearchDatastore_Task, 'duration_secs': 0.00892} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.132145] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.132145] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.132293] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.132585] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.132672] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.132974] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e30c2e70-459b-4222-92ea-edd5cdbc3fb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.141984] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.142200] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.142945] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a94ee00-cf0a-4056-892a-f2124ed0d4a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.148439] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 985.148439] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5283e284-d82a-9591-f67b-d2c1d121de68" [ 985.148439] env[63175]: _type = "Task" [ 985.148439] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.156934] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5283e284-d82a-9591-f67b-d2c1d121de68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.197634] env[63175]: DEBUG nova.scheduler.client.report [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 985.477020] env[63175]: DEBUG nova.compute.manager [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 985.477327] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.478504] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55ef661-166d-4ae0-bdec-74ab219173da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.486379] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.486647] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5ae71e4-c4e3-4dd8-954e-920433079f25 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.492930] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 985.492930] env[63175]: value = "task-1248459" [ 985.492930] env[63175]: _type = "Task" [ 985.492930] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.500679] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.658646] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5283e284-d82a-9591-f67b-d2c1d121de68, 'name': SearchDatastore_Task, 'duration_secs': 0.008826} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.659485] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-798421cb-ccf0-4f3d-9451-3edbb4919c7b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.665787] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 985.665787] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5205aeea-3bb4-f044-fbfd-50794afd14cd" [ 985.665787] env[63175]: _type = "Task" [ 985.665787] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.673864] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5205aeea-3bb4-f044-fbfd-50794afd14cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.702283] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.723880] env[63175]: INFO nova.scheduler.client.report [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted allocations for instance 63be6b3c-2535-4c26-9a15-9c2445b049e8 [ 986.003227] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248459, 'name': PowerOffVM_Task, 'duration_secs': 0.201237} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.003547] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.003804] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.004187] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86d13329-fb00-4ea7-a5a1-6939bf520c94 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.012797] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 986.038954] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.039241] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.039406] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.039639] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.039798] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.039953] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.040185] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.040354] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.040557] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.040686] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.040942] env[63175]: DEBUG nova.virt.hardware [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.041825] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887eb27f-1227-4250-8fb2-bb268657247f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.049659] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dff84d1-17a0-49d9-8dfe-c15c3ecc129b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.071091] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.071296] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.071486] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleting the datastore file [datastore1] 0645fffe-9910-43af-af40-126592aefaff {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.071748] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6126c6f-6c0d-41e9-859f-6e9b7c7d906e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.078136] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 986.078136] env[63175]: value = "task-1248461" [ 986.078136] env[63175]: _type = "Task" [ 986.078136] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.085935] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.176812] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5205aeea-3bb4-f044-fbfd-50794afd14cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.177140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.177461] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0/9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.178332] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdc160e1-d99a-43a2-814f-cc93ec0a926f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.184946] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 986.184946] env[63175]: value = "task-1248462" [ 986.184946] env[63175]: _type = "Task" [ 986.184946] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.192334] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.233386] env[63175]: DEBUG oslo_concurrency.lockutils [None req-75b10620-d3a4-4a9b-8533-a1979d65317f tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "63be6b3c-2535-4c26-9a15-9c2445b049e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.977s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.473626] env[63175]: DEBUG nova.compute.manager [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-vif-plugged-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 986.473857] env[63175]: DEBUG oslo_concurrency.lockutils [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.474117] env[63175]: DEBUG oslo_concurrency.lockutils [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.474303] env[63175]: DEBUG oslo_concurrency.lockutils [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.474466] env[63175]: DEBUG nova.compute.manager [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] No waiting events found dispatching network-vif-plugged-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 986.474645] env[63175]: WARNING nova.compute.manager [req-a64bc576-a670-46bd-a164-976391793e34 req-4e30660b-76fe-47be-9f82-09a4139b8d50 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received unexpected event network-vif-plugged-298629e3-8cd5-4c3d-89e7-c008ca345f71 for instance with vm_state building and task_state spawning. [ 986.594418] env[63175]: DEBUG oslo_vmware.api [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133789} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.594623] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.594855] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.595243] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.595637] env[63175]: INFO nova.compute.manager [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 0645fffe-9910-43af-af40-126592aefaff] Took 1.12 seconds to destroy the instance on the hypervisor. [ 986.595949] env[63175]: DEBUG oslo.service.loopingcall [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.596356] env[63175]: DEBUG nova.compute.manager [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 986.596512] env[63175]: DEBUG nova.network.neutron [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.695488] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499905} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.697608] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0/9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.697839] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.698299] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4aafb2eb-4d59-4d0c-8a56-ec5190ce2e09 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.705211] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 986.705211] env[63175]: value = "task-1248463" [ 986.705211] env[63175]: _type = "Task" [ 986.705211] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.722469] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.781109] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Successfully updated port: 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.910026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.910026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.910026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.910026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.910026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.912646] env[63175]: INFO nova.compute.manager [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Terminating instance [ 987.215910] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067318} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.215910] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 987.216572] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90e9407-3b89-481e-9acf-aec77181f2a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.248927] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0/9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 987.249591] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-949ae4de-6f47-4a62-a94e-11c83ccaea3c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.273008] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 987.273008] env[63175]: value = "task-1248464" [ 987.273008] env[63175]: _type = "Task" [ 987.273008] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.280805] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.285466] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.285599] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.285722] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.418058] env[63175]: DEBUG nova.compute.manager [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 987.418288] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.419253] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f13e71d-33c4-443d-81c7-c1c63434a8bb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.427412] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.427697] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-240d98ab-221f-44d0-8930-70c3cd386e8e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.434324] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 987.434324] env[63175]: value = "task-1248465" [ 987.434324] env[63175]: _type = "Task" [ 987.434324] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.443845] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248465, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.732653] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.732894] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.783387] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248464, 'name': ReconfigVM_Task, 'duration_secs': 0.295788} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.783792] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0/9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.784600] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-307a79a2-cb4e-4ffb-a9a9-67ef487f1093 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.791759] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 987.791759] env[63175]: value = "task-1248466" [ 987.791759] env[63175]: _type = "Task" [ 987.791759] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.802258] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248466, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.828775] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.871366] env[63175]: DEBUG nova.network.neutron [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.956575] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248465, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.985875] env[63175]: DEBUG nova.network.neutron [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.235780] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 988.301929] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248466, 'name': Rename_Task, 'duration_secs': 0.139468} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.302586] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.302719] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71e69658-02c0-462c-9cff-47474123c96e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.309288] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 988.309288] env[63175]: value = "task-1248467" [ 988.309288] env[63175]: _type = "Task" [ 988.309288] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.316914] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248467, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.375145] env[63175]: INFO nova.compute.manager [-] [instance: 0645fffe-9910-43af-af40-126592aefaff] Took 1.78 seconds to deallocate network for instance. [ 988.446281] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248465, 'name': PowerOffVM_Task, 'duration_secs': 0.565415} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.446479] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.450019] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.450019] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ee53188-cbc1-4474-9d87-608d42aa7ea5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.490095] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.490473] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Instance network_info: |[{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 988.490944] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:de:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '298629e3-8cd5-4c3d-89e7-c008ca345f71', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.499218] env[63175]: DEBUG oslo.service.loopingcall [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.500583] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.502614] env[63175]: DEBUG nova.compute.manager [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 988.502831] env[63175]: DEBUG nova.compute.manager [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing instance network info cache due to event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 988.502971] env[63175]: DEBUG oslo_concurrency.lockutils [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.503137] env[63175]: DEBUG oslo_concurrency.lockutils [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.503302] env[63175]: DEBUG nova.network.neutron [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.504637] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cdedad61-9d8b-4635-a894-40f7fa7329a2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.524195] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.524432] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.524618] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore2] 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.525551] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0165212b-a858-4bba-a5d8-5ca5a1834296 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.530024] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.530024] env[63175]: value = "task-1248469" [ 988.530024] env[63175]: _type = "Task" [ 988.530024] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.531178] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 988.531178] env[63175]: value = "task-1248470" [ 988.531178] env[63175]: _type = "Task" [ 988.531178] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.545483] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248469, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.548588] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.750338] env[63175]: DEBUG nova.network.neutron [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updated VIF entry in instance network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.750693] env[63175]: DEBUG nova.network.neutron [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.767322] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.767591] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.769647] env[63175]: INFO nova.compute.claims [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.821318] env[63175]: DEBUG oslo_vmware.api [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248467, 'name': PowerOnVM_Task, 'duration_secs': 0.449951} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.821607] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.821826] env[63175]: INFO nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Took 7.47 seconds to spawn the instance on the hypervisor. [ 988.821988] env[63175]: DEBUG nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 988.823047] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd519125-d231-49eb-8847-c95b319b85c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.881809] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.048935] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248469, 'name': CreateVM_Task, 'duration_secs': 0.3605} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.051958] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.052280] env[63175]: DEBUG oslo_vmware.api [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157021} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.052943] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.053234] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.053572] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.053826] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.054101] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.054299] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.055098] env[63175]: INFO nova.compute.manager [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Took 1.64 seconds to destroy the instance on the hypervisor. [ 989.055796] env[63175]: DEBUG oslo.service.loopingcall [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.056039] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed68797e-1d87-40ff-9224-7c6eb5565262 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.057670] env[63175]: DEBUG nova.compute.manager [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 989.057779] env[63175]: DEBUG nova.network.neutron [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.062321] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 989.062321] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3af31-7ced-cce1-c2ff-558dfc6e58fe" [ 989.062321] env[63175]: _type = "Task" [ 989.062321] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.070153] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3af31-7ced-cce1-c2ff-558dfc6e58fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.256029] env[63175]: DEBUG oslo_concurrency.lockutils [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.256029] env[63175]: DEBUG nova.compute.manager [req-403bfdc9-22b6-42c8-9a6e-5a38e2b2f4dc req-30bd6678-8dd6-49be-9e01-5bd87710a325 service nova] [instance: 0645fffe-9910-43af-af40-126592aefaff] Received event network-vif-deleted-a16e4e4f-df78-450a-bf4d-d43daac38791 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 989.341621] env[63175]: INFO nova.compute.manager [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Took 13.20 seconds to build instance. [ 989.572162] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d3af31-7ced-cce1-c2ff-558dfc6e58fe, 'name': SearchDatastore_Task, 'duration_secs': 0.016594} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.572584] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.572887] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.573163] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.573338] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.573533] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.573807] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99fc8398-dc84-4058-b8e9-1e528dc87cd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.582360] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.582545] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.583270] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39107f84-e4b5-4b62-9725-0b3345588bd0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.588181] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 989.588181] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bc904-aa86-2a43-e223-c691ec924573" [ 989.588181] env[63175]: _type = "Task" [ 989.588181] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.595851] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bc904-aa86-2a43-e223-c691ec924573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.843313] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9ae55fa9-ad2c-40ab-9f40-f1210e213ea7 tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.710s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.941731] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6565f9d-c23d-47a6-847c-8a5661b64738 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.952220] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6445941-f5d5-478f-a4e9-e9c03f21dab3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.982356] env[63175]: DEBUG nova.network.neutron [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.984246] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980aeac6-284b-4ce6-bf70-1c804484813c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.993536] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d932ea-7b0b-4ba4-8fac-3fa0accbb567 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.010160] env[63175]: DEBUG nova.compute.provider_tree [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.018329] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.018652] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.018778] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.018987] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.019140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.023374] env[63175]: INFO nova.compute.manager [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Terminating instance [ 990.098277] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529bc904-aa86-2a43-e223-c691ec924573, 'name': SearchDatastore_Task, 'duration_secs': 0.008414} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.098945] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6316f0e3-f1f5-4b8f-aac5-6729a01db4b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.104576] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 990.104576] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f00fbd-aa45-be1a-1081-6d98e0bb0702" [ 990.104576] env[63175]: _type = "Task" [ 990.104576] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.111568] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f00fbd-aa45-be1a-1081-6d98e0bb0702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.488016] env[63175]: INFO nova.compute.manager [-] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Took 1.43 seconds to deallocate network for instance. [ 990.514006] env[63175]: DEBUG nova.scheduler.client.report [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 990.527230] env[63175]: DEBUG nova.compute.manager [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 990.527474] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.529613] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de046650-0f92-44f1-aef1-f0911911c745 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.533935] env[63175]: DEBUG nova.compute.manager [req-d8201061-967f-497c-92f0-61171247b98a req-63d72cf5-ca8c-43fd-a002-3619d8f9c59e service nova] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Received event network-vif-deleted-398121e4-74d6-495f-a429-da631d9be9a5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 990.540426] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.540714] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb34d3e1-d5d7-498e-880f-2b9f7ccabbc5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.547511] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 990.547511] env[63175]: value = "task-1248471" [ 990.547511] env[63175]: _type = "Task" [ 990.547511] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.556454] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.615515] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f00fbd-aa45-be1a-1081-6d98e0bb0702, 'name': SearchDatastore_Task, 'duration_secs': 0.010338} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.615823] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.616135] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 8a4877b2-6284-4727-b1fc-cf30bbc65e36/8a4877b2-6284-4727-b1fc-cf30bbc65e36.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.616430] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9aaa410-8fb0-472b-b02f-65e39b054ecc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.624487] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 990.624487] env[63175]: value = "task-1248472" [ 990.624487] env[63175]: _type = "Task" [ 990.624487] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.632912] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.995224] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.019813] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.020501] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 991.023914] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.142s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.023914] env[63175]: DEBUG nova.objects.instance [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'resources' on Instance uuid 0645fffe-9910-43af-af40-126592aefaff {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.061954] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248471, 'name': PowerOffVM_Task, 'duration_secs': 0.194298} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.063209] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.063209] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.063209] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e06e13a3-c186-437e-9f73-955e0c995095 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.139349] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.139707] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.140259] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Deleting the datastore file [datastore2] 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.140259] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248472, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.140626] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0525ea28-a23f-4c18-8148-61d8702a8a48 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.147981] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for the task: (returnval){ [ 991.147981] env[63175]: value = "task-1248474" [ 991.147981] env[63175]: _type = "Task" [ 991.147981] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.159185] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.527276] env[63175]: DEBUG nova.compute.utils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.532072] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 991.532257] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 991.574525] env[63175]: DEBUG nova.policy [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e43774590162406a981cf0d5e20f5a1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83fa8f008a48484195c88ed76ceece0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.638083] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660566} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.638397] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 8a4877b2-6284-4727-b1fc-cf30bbc65e36/8a4877b2-6284-4727-b1fc-cf30bbc65e36.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.638646] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.638883] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2cd1eb76-2b0d-4ddd-980a-d823a5658d67 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.646181] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 991.646181] env[63175]: value = "task-1248475" [ 991.646181] env[63175]: _type = "Task" [ 991.646181] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.658351] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248475, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.664345] env[63175]: DEBUG oslo_vmware.api [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Task: {'id': task-1248474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259303} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.664869] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.665367] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.665367] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.665523] env[63175]: INFO nova.compute.manager [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 991.665727] env[63175]: DEBUG oslo.service.loopingcall [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.665954] env[63175]: DEBUG nova.compute.manager [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 991.666029] env[63175]: DEBUG nova.network.neutron [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 991.695580] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ad9ff8-c15d-4f71-8afc-cc98f7192d2a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.704147] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eac546-4324-4d9e-8da1-c4e29ccdf3f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.742407] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903b6e45-c49d-41dd-8693-50a53af3fa9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.748230] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "19ec0aa6-3078-419d-849d-1bb371325a87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.748473] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.754489] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a43912-0b89-42c2-8e1f-def985dfd1ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.769293] env[63175]: DEBUG nova.compute.provider_tree [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.037613] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 992.158351] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248475, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089583} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.158351] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.159185] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8199efaf-522c-4c0e-97c6-d0bf85512b45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.186251] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 8a4877b2-6284-4727-b1fc-cf30bbc65e36/8a4877b2-6284-4727-b1fc-cf30bbc65e36.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.186786] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f3d775e-3162-4cdd-853d-cb9650456b7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.206347] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Successfully created port: ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.217026] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 992.217026] env[63175]: value = "task-1248476" [ 992.217026] env[63175]: _type = "Task" [ 992.217026] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.224511] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248476, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.254298] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 992.273606] env[63175]: DEBUG nova.scheduler.client.report [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 992.559445] env[63175]: DEBUG nova.compute.manager [req-011d5a29-5710-41c1-8917-dca5eb118a92 req-6a7bb38b-4254-42a0-b18e-722332cabcce service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Received event network-vif-deleted-34325eb5-e47f-446d-925c-bc3a7e76a2d3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 992.559445] env[63175]: INFO nova.compute.manager [req-011d5a29-5710-41c1-8917-dca5eb118a92 req-6a7bb38b-4254-42a0-b18e-722332cabcce service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Neutron deleted interface 34325eb5-e47f-446d-925c-bc3a7e76a2d3; detaching it from the instance and deleting it from the info cache [ 992.559445] env[63175]: DEBUG nova.network.neutron [req-011d5a29-5710-41c1-8917-dca5eb118a92 req-6a7bb38b-4254-42a0-b18e-722332cabcce service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.725142] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248476, 'name': ReconfigVM_Task, 'duration_secs': 0.327349} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.725435] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 8a4877b2-6284-4727-b1fc-cf30bbc65e36/8a4877b2-6284-4727-b1fc-cf30bbc65e36.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.726095] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4f35ba4-353f-4181-a8de-89b7b727e2d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.733173] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 992.733173] env[63175]: value = "task-1248477" [ 992.733173] env[63175]: _type = "Task" [ 992.733173] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.736562] env[63175]: DEBUG nova.network.neutron [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.740808] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248477, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.774438] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.779611] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.782504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.787s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.782504] env[63175]: DEBUG nova.objects.instance [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'resources' on Instance uuid 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.799353] env[63175]: INFO nova.scheduler.client.report [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted allocations for instance 0645fffe-9910-43af-af40-126592aefaff [ 993.045425] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 993.061547] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e9786e8-9c4d-4d16-b7db-97cba7a95630 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.071604] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.071845] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.072014] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.072214] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.072366] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.072517] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.072748] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.072929] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.073119] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.073289] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.073466] env[63175]: DEBUG nova.virt.hardware [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.076330] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b978587a-f6fb-470f-a625-8f792c4ba79a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.086836] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0510de07-fc71-41ff-843f-823b931f8509 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.094288] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae8aaf6-0a49-4f44-84a2-b7194634455b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.106050] env[63175]: DEBUG nova.compute.manager [req-011d5a29-5710-41c1-8917-dca5eb118a92 req-6a7bb38b-4254-42a0-b18e-722332cabcce service nova] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Detach interface failed, port_id=34325eb5-e47f-446d-925c-bc3a7e76a2d3, reason: Instance 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 993.242928] env[63175]: INFO nova.compute.manager [-] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Took 1.58 seconds to deallocate network for instance. [ 993.243282] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248477, 'name': Rename_Task, 'duration_secs': 0.148279} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.244983] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.247617] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b301b9b8-a7d4-4693-881d-93e691dd19df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.254220] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 993.254220] env[63175]: value = "task-1248478" [ 993.254220] env[63175]: _type = "Task" [ 993.254220] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.261889] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.310086] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8fb0f89d-73f8-4ea3-b353-29aa0ab8a49b tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "0645fffe-9910-43af-af40-126592aefaff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.341s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.425446] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a523111d-d8a2-492c-9338-91efe085fe9e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.434435] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a490b18-ee58-4449-a86b-650d845631e6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.466584] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a686de42-29cc-4844-b802-0f42e70f631f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.474722] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5de0e68-0755-4e8a-a1f5-edcbb38fce2c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.488357] env[63175]: DEBUG nova.compute.provider_tree [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.753287] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.768215] env[63175]: DEBUG oslo_vmware.api [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248478, 'name': PowerOnVM_Task, 'duration_secs': 0.446252} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.768215] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.768215] env[63175]: INFO nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Took 7.75 seconds to spawn the instance on the hypervisor. [ 993.768215] env[63175]: DEBUG nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 993.768215] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38015d3-4de2-45ae-9e4a-4faa697f29b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.819518] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Successfully updated port: ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.992318] env[63175]: DEBUG nova.scheduler.client.report [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 994.814941] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.815304] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.815304] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.816548] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.035s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.819238] env[63175]: DEBUG nova.compute.manager [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Received event network-vif-plugged-ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 994.819426] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Acquiring lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.819631] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.819794] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.819955] env[63175]: DEBUG nova.compute.manager [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] No waiting events found dispatching network-vif-plugged-ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 994.820132] env[63175]: WARNING nova.compute.manager [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Received unexpected event network-vif-plugged-ed916caf-e13e-4179-9349-feabcfaa75b3 for instance with vm_state building and task_state spawning. [ 994.820286] env[63175]: DEBUG nova.compute.manager [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Received event network-changed-ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 994.820434] env[63175]: DEBUG nova.compute.manager [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Refreshing instance network info cache due to event network-changed-ed916caf-e13e-4179-9349-feabcfaa75b3. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 994.820600] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Acquiring lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.822829] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.049s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.824950] env[63175]: INFO nova.compute.claims [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.827704] env[63175]: INFO nova.compute.manager [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Took 13.19 seconds to build instance. [ 994.851658] env[63175]: INFO nova.scheduler.client.report [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted allocations for instance 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db [ 995.240088] env[63175]: DEBUG nova.compute.manager [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 995.240088] env[63175]: DEBUG nova.compute.manager [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 995.240738] env[63175]: DEBUG oslo_concurrency.lockutils [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.240738] env[63175]: DEBUG oslo_concurrency.lockutils [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.240982] env[63175]: DEBUG nova.network.neutron [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.334069] env[63175]: DEBUG oslo_concurrency.lockutils [None req-62de053c-89d7-423d-8f58-a25ef13f73eb tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.700s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.360282] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.362336] env[63175]: DEBUG oslo_concurrency.lockutils [None req-560a7e77-895b-4b35-9ff8-e2a21aaafba6 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.453s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.573692] env[63175]: DEBUG nova.network.neutron [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Updating instance_info_cache with network_info: [{"id": "ed916caf-e13e-4179-9349-feabcfaa75b3", "address": "fa:16:3e:ea:05:f3", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped916caf-e1", "ovs_interfaceid": "ed916caf-e13e-4179-9349-feabcfaa75b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.956259] env[63175]: DEBUG nova.network.neutron [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.956584] env[63175]: DEBUG nova.network.neutron [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.985860] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f893ab16-bdb8-4e81-9acc-7b627b2482ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.995033] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd5ed4a-a875-471d-97ba-57a3e9752e30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.027074] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300fa369-ac2c-4904-89bd-cdb823d4e4ae {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.034658] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3408f721-7c3d-4050-91b4-d8062d27f42b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.048243] env[63175]: DEBUG nova.compute.provider_tree [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.074845] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.075169] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Instance network_info: |[{"id": "ed916caf-e13e-4179-9349-feabcfaa75b3", "address": "fa:16:3e:ea:05:f3", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped916caf-e1", "ovs_interfaceid": "ed916caf-e13e-4179-9349-feabcfaa75b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 996.075507] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Acquired lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.075721] env[63175]: DEBUG nova.network.neutron [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Refreshing network info cache for port ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.077029] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:05:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '721c6720-3ce0-450e-9951-a894f03acc27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed916caf-e13e-4179-9349-feabcfaa75b3', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.085217] env[63175]: DEBUG oslo.service.loopingcall [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.088763] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 996.089215] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b219d1a3-6ab6-4f91-94e5-ccac3fb7f29e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.109953] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.109953] env[63175]: value = "task-1248480" [ 996.109953] env[63175]: _type = "Task" [ 996.109953] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.117430] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248480, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.370886] env[63175]: DEBUG nova.network.neutron [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Updated VIF entry in instance network info cache for port ed916caf-e13e-4179-9349-feabcfaa75b3. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 996.371340] env[63175]: DEBUG nova.network.neutron [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Updating instance_info_cache with network_info: [{"id": "ed916caf-e13e-4179-9349-feabcfaa75b3", "address": "fa:16:3e:ea:05:f3", "network": {"id": "2986958e-ae9f-4647-b21e-a1220de568b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1910679381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83fa8f008a48484195c88ed76ceece0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "721c6720-3ce0-450e-9951-a894f03acc27", "external-id": "nsx-vlan-transportzone-394", "segmentation_id": 394, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped916caf-e1", "ovs_interfaceid": "ed916caf-e13e-4179-9349-feabcfaa75b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.459687] env[63175]: DEBUG oslo_concurrency.lockutils [req-3a687174-baf8-4442-aa96-24e9b2a6d5d4 req-a8b3cd96-6972-4da5-983c-736e27b64890 service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.550621] env[63175]: DEBUG nova.scheduler.client.report [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 996.623369] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248480, 'name': CreateVM_Task, 'duration_secs': 0.315232} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.623369] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.624056] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.624412] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.624673] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.624938] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e97e5acc-7b7a-4b37-979e-b91ff1a114dc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.629457] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 996.629457] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5265cd25-cfe7-3d26-970a-6c5aca16806f" [ 996.629457] env[63175]: _type = "Task" [ 996.629457] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.637185] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5265cd25-cfe7-3d26-970a-6c5aca16806f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.874531] env[63175]: DEBUG oslo_concurrency.lockutils [req-184577a3-dd30-451b-9e1f-7ac5cf651110 req-2ebc60e0-100c-4711-8800-4ab95445a451 service nova] Releasing lock "refresh_cache-b8c2febb-7024-49de-b34c-1c8ee492b39f" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.055691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.056264] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 997.059066] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.306s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.059302] env[63175]: DEBUG nova.objects.instance [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lazy-loading 'resources' on Instance uuid 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.141762] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5265cd25-cfe7-3d26-970a-6c5aca16806f, 'name': SearchDatastore_Task, 'duration_secs': 0.009746} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.142113] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.142351] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.142589] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.142742] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.142992] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.143191] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78b60685-2fd1-475f-a52f-6f91a7c1aa0d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.157236] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.157514] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.158178] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4440cae-12cd-404f-bf6d-9d490e85256c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.163520] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 997.163520] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5212d414-e3a9-3831-b698-5f85675851c2" [ 997.163520] env[63175]: _type = "Task" [ 997.163520] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.171537] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5212d414-e3a9-3831-b698-5f85675851c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.268855] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 997.268855] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing instance network info cache due to event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 997.268855] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.268971] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.269101] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.530559] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.530840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.532051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.532051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.532051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.533664] env[63175]: INFO nova.compute.manager [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Terminating instance [ 997.561507] env[63175]: DEBUG nova.compute.utils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.565752] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 997.565887] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 997.603498] env[63175]: DEBUG nova.policy [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba1ccac5ed624dabb1046c8be095d389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac46649a931423598655c2f550ce7b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.673530] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5212d414-e3a9-3831-b698-5f85675851c2, 'name': SearchDatastore_Task, 'duration_secs': 0.026358} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.676666] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2c3bf58-8285-4ec4-b3df-096a711ec4f6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.681534] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 997.681534] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52444bec-d85d-27c6-78d8-d505f951198b" [ 997.681534] env[63175]: _type = "Task" [ 997.681534] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.686980] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c34ada7-58b0-40bf-850e-3cd1eeab461b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.691882] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52444bec-d85d-27c6-78d8-d505f951198b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.696176] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cac653-1538-4ac6-b426-87cb61530c61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.725530] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3536332b-fbca-4892-9583-82efc3082293 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.732495] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd400e1-aa7d-4dd4-85e9-72e27ad22a72 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.744981] env[63175]: DEBUG nova.compute.provider_tree [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.931976] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Successfully created port: 1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.037146] env[63175]: DEBUG nova.compute.manager [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 998.037372] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.038234] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6d04c7-cec2-408b-94c1-982ab06c3a35 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.045898] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.046142] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b533d149-94fc-4d19-afe6-956775fa7c14 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.052083] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 998.052083] env[63175]: value = "task-1248481" [ 998.052083] env[63175]: _type = "Task" [ 998.052083] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.059671] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.066296] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 998.083268] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updated VIF entry in instance network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.083668] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.191881] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52444bec-d85d-27c6-78d8-d505f951198b, 'name': SearchDatastore_Task, 'duration_secs': 0.016951} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.192182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.192580] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] b8c2febb-7024-49de-b34c-1c8ee492b39f/b8c2febb-7024-49de-b34c-1c8ee492b39f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.192712] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b50ffda9-a9ca-4171-802e-5f5f82272f41 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.199072] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 998.199072] env[63175]: value = "task-1248482" [ 998.199072] env[63175]: _type = "Task" [ 998.199072] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.207727] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.212790] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.213029] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.247792] env[63175]: DEBUG nova.scheduler.client.report [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 998.562216] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248481, 'name': PowerOffVM_Task, 'duration_secs': 0.263476} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.562538] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.562746] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.563039] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea9d8da9-ed02-45c0-b28f-59fd26559a16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.586819] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.587148] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 998.587331] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing instance network info cache due to event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 998.587555] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.587704] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.587941] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 998.649597] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.649856] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.649997] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleting the datastore file [datastore1] 54440032-4d9b-41d4-9ef2-5a79a4224fa6 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.650287] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cdbd1a2-4b92-4aec-a81c-2364033dc458 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.655894] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 998.655894] env[63175]: value = "task-1248484" [ 998.655894] env[63175]: _type = "Task" [ 998.655894] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.664224] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.708258] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43718} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.708512] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] b8c2febb-7024-49de-b34c-1c8ee492b39f/b8c2febb-7024-49de-b34c-1c8ee492b39f.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.708776] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.709085] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d29725b2-d4ff-4246-b5bf-2c649dcb603b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.716389] env[63175]: DEBUG nova.compute.utils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.719553] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 998.719553] env[63175]: value = "task-1248485" [ 998.719553] env[63175]: _type = "Task" [ 998.719553] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.727164] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248485, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.753504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.776289] env[63175]: INFO nova.scheduler.client.report [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Deleted allocations for instance 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0 [ 999.076270] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 999.102609] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.102937] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.103156] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.103360] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.103518] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.103672] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.103883] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.104063] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.104246] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.104489] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.104736] env[63175]: DEBUG nova.virt.hardware [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.105692] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abc8f68-4859-4ed5-8662-0479495e6d7d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.114281] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a506e9-c1b5-4a4b-a903-d2453e349952 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.166649] env[63175]: DEBUG oslo_vmware.api [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169235} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.166969] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.167169] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.167362] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.167543] env[63175]: INFO nova.compute.manager [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Took 1.13 seconds to destroy the instance on the hypervisor. [ 999.167823] env[63175]: DEBUG oslo.service.loopingcall [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.167998] env[63175]: DEBUG nova.compute.manager [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 999.168104] env[63175]: DEBUG nova.network.neutron [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.219535] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.229139] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248485, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066072} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.229394] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.230176] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e03800-ae7e-4e7b-98a7-10d5f6992359 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.254070] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] b8c2febb-7024-49de-b34c-1c8ee492b39f/b8c2febb-7024-49de-b34c-1c8ee492b39f.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.254353] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2d01623-3144-4141-a439-e07998e48714 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.273576] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 999.273576] env[63175]: value = "task-1248486" [ 999.273576] env[63175]: _type = "Task" [ 999.273576] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.283576] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248486, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.284043] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f6983357-1139-4b54-9df1-a6d8e2ba5eea tempest-ServerPasswordTestJSON-107560156 tempest-ServerPasswordTestJSON-107560156-project-member] Lock "9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.265s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.368084] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updated VIF entry in instance network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.368522] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.476022] env[63175]: DEBUG nova.compute.manager [req-35a70520-ba64-4d16-bc81-d3644cd223f9 req-6ac2ab48-0304-43b0-a916-b8126fb82aa6 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Received event network-vif-deleted-8ec148c2-7cf8-4923-9c17-a6f88820cf10 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.476274] env[63175]: INFO nova.compute.manager [req-35a70520-ba64-4d16-bc81-d3644cd223f9 req-6ac2ab48-0304-43b0-a916-b8126fb82aa6 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Neutron deleted interface 8ec148c2-7cf8-4923-9c17-a6f88820cf10; detaching it from the instance and deleting it from the info cache [ 999.476423] env[63175]: DEBUG nova.network.neutron [req-35a70520-ba64-4d16-bc81-d3644cd223f9 req-6ac2ab48-0304-43b0-a916-b8126fb82aa6 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.496067] env[63175]: DEBUG nova.compute.manager [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Received event network-vif-plugged-1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.496307] env[63175]: DEBUG oslo_concurrency.lockutils [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] Acquiring lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.496523] env[63175]: DEBUG oslo_concurrency.lockutils [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] Lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.496719] env[63175]: DEBUG oslo_concurrency.lockutils [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] Lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.496876] env[63175]: DEBUG nova.compute.manager [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] No waiting events found dispatching network-vif-plugged-1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 999.497136] env[63175]: WARNING nova.compute.manager [req-9dafbce3-3b35-489e-8d03-8e894d686d14 req-e426f563-073f-4a51-8753-89b2627c79a2 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Received unexpected event network-vif-plugged-1aa15971-a833-49bd-ae85-64dbc5908057 for instance with vm_state building and task_state spawning. [ 999.560528] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Successfully updated port: 1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.788244] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248486, 'name': ReconfigVM_Task, 'duration_secs': 0.312791} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.788687] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Reconfigured VM instance instance-0000005c to attach disk [datastore1] b8c2febb-7024-49de-b34c-1c8ee492b39f/b8c2febb-7024-49de-b34c-1c8ee492b39f.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.789372] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a581898-fb7d-4dbc-86d5-b2dd80297216 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.798688] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 999.798688] env[63175]: value = "task-1248487" [ 999.798688] env[63175]: _type = "Task" [ 999.798688] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.809694] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248487, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.871045] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.871302] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.871490] env[63175]: DEBUG nova.compute.manager [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 999.871903] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.872172] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.872282] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.910716] env[63175]: DEBUG nova.network.neutron [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.980240] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c64bb272-e281-487a-8e43-1857d58c574f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.989972] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463c087f-99f4-492e-9cda-3ef7350193a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.015069] env[63175]: DEBUG nova.compute.manager [req-35a70520-ba64-4d16-bc81-d3644cd223f9 req-6ac2ab48-0304-43b0-a916-b8126fb82aa6 service nova] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Detach interface failed, port_id=8ec148c2-7cf8-4923-9c17-a6f88820cf10, reason: Instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1000.062990] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.062990] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.063161] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.299868] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.300247] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.300507] env[63175]: INFO nova.compute.manager [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Attaching volume a02a3124-f4cb-440b-bde1-d22a3407142d to /dev/sdb [ 1000.315288] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248487, 'name': Rename_Task, 'duration_secs': 0.132228} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.315576] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.315832] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9f0a869-60b8-462e-b796-242bcb01cb70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.321479] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1000.321479] env[63175]: value = "task-1248488" [ 1000.321479] env[63175]: _type = "Task" [ 1000.321479] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.332833] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248488, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.345847] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07817e3-168f-4fc7-a4a9-0f95ad82977f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.358021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4822ed49-b5b9-4bf7-9c80-237f92722c7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.372635] env[63175]: DEBUG nova.virt.block_device [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updating existing volume attachment record: 8ea47601-20a1-4a34-8f25-c36dd35d95b1 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1000.415153] env[63175]: INFO nova.compute.manager [-] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Took 1.25 seconds to deallocate network for instance. [ 1000.602050] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.737315] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.737722] env[63175]: DEBUG nova.network.neutron [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.740103] env[63175]: DEBUG nova.network.neutron [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updating instance_info_cache with network_info: [{"id": "1aa15971-a833-49bd-ae85-64dbc5908057", "address": "fa:16:3e:36:52:49", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa15971-a8", "ovs_interfaceid": "1aa15971-a833-49bd-ae85-64dbc5908057", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.832398] env[63175]: DEBUG oslo_vmware.api [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248488, 'name': PowerOnVM_Task, 'duration_secs': 0.491077} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.832708] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.832968] env[63175]: INFO nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Took 7.79 seconds to spawn the instance on the hypervisor. [ 1000.833134] env[63175]: DEBUG nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1000.833911] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cb42f7-7fda-4b26-a0e0-43d8cdd3cb8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.924954] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.925654] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.925654] env[63175]: DEBUG nova.objects.instance [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'resources' on Instance uuid 54440032-4d9b-41d4-9ef2-5a79a4224fa6 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.240287] env[63175]: DEBUG oslo_concurrency.lockutils [req-5dca337d-d5d4-44b6-8d6b-a5c8ec22c6ca req-d2d401c9-5b70-4f18-835b-6c373f6c507a service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.243085] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.243436] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Instance network_info: |[{"id": "1aa15971-a833-49bd-ae85-64dbc5908057", "address": "fa:16:3e:36:52:49", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa15971-a8", "ovs_interfaceid": "1aa15971-a833-49bd-ae85-64dbc5908057", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1001.243866] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:52:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f246b87-f105-4b33-a71d-5caf8e99e074', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1aa15971-a833-49bd-ae85-64dbc5908057', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.251495] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating folder: Project (eac46649a931423598655c2f550ce7b4). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1001.251797] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1d7bf98-06b3-4019-9eff-c1a58a2f0383 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.263373] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created folder: Project (eac46649a931423598655c2f550ce7b4) in parent group-v268956. [ 1001.263597] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating folder: Instances. Parent ref: group-v269088. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1001.263878] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7c9a421-1049-428d-8851-78cca7e7b73b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.272677] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created folder: Instances in parent group-v269088. [ 1001.272965] env[63175]: DEBUG oslo.service.loopingcall [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.273249] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.273544] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0796647-8040-4ec7-adb1-e4e3404ac600 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.295117] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.295117] env[63175]: value = "task-1248492" [ 1001.295117] env[63175]: _type = "Task" [ 1001.295117] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.303156] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248492, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.352896] env[63175]: INFO nova.compute.manager [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Took 12.61 seconds to build instance. [ 1001.529178] env[63175]: DEBUG nova.compute.manager [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Received event network-changed-1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1001.529178] env[63175]: DEBUG nova.compute.manager [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Refreshing instance network info cache due to event network-changed-1aa15971-a833-49bd-ae85-64dbc5908057. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1001.529771] env[63175]: DEBUG oslo_concurrency.lockutils [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] Acquiring lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.529771] env[63175]: DEBUG oslo_concurrency.lockutils [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] Acquired lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.529771] env[63175]: DEBUG nova.network.neutron [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Refreshing network info cache for port 1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.547253] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4cdd6c-b46b-466a-881e-33f23ac3c5f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.555264] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef44f5d-6829-407a-9b11-c4c4e3fb988b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.587722] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7804d115-320f-4169-9ac1-df846e35ca0f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.595164] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3cf70b-4a62-4a7d-8432-9feb66f2fcfa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.608728] env[63175]: DEBUG nova.compute.provider_tree [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.805843] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248492, 'name': CreateVM_Task, 'duration_secs': 0.344889} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.806150] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.807009] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.807223] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.807551] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.807815] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-800a4cc5-c6a7-49d2-9061-08b2aca70a36 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.812782] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1001.812782] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f38b2c-cbd8-c930-1361-7375df377936" [ 1001.812782] env[63175]: _type = "Task" [ 1001.812782] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.820964] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f38b2c-cbd8-c930-1361-7375df377936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.854992] env[63175]: DEBUG oslo_concurrency.lockutils [None req-9029bb89-27da-4272-aa58-3c0072d5cd35 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.122s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.113050] env[63175]: DEBUG nova.scheduler.client.report [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1002.322682] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52f38b2c-cbd8-c930-1361-7375df377936, 'name': SearchDatastore_Task, 'duration_secs': 0.025543} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.323049] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.323217] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.323454] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.323634] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.323857] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.324133] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-268d7540-110d-4899-89b7-76608e979228 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.332397] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.332576] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.333270] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-689c6779-46c9-45ac-88d0-8cd7dd2b560e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.337991] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1002.337991] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fd1fd9-c3b9-9e7e-573b-296e4bed9945" [ 1002.337991] env[63175]: _type = "Task" [ 1002.337991] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.349519] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fd1fd9-c3b9-9e7e-573b-296e4bed9945, 'name': SearchDatastore_Task, 'duration_secs': 0.008375} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.350466] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e6b26da-e254-407e-9aa8-5146588f3221 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.354925] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1002.354925] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52380f29-056e-035c-ce46-c7c14b42f83e" [ 1002.354925] env[63175]: _type = "Task" [ 1002.354925] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.363061] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52380f29-056e-035c-ce46-c7c14b42f83e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.435267] env[63175]: DEBUG nova.network.neutron [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updated VIF entry in instance network info cache for port 1aa15971-a833-49bd-ae85-64dbc5908057. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.435868] env[63175]: DEBUG nova.network.neutron [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updating instance_info_cache with network_info: [{"id": "1aa15971-a833-49bd-ae85-64dbc5908057", "address": "fa:16:3e:36:52:49", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa15971-a8", "ovs_interfaceid": "1aa15971-a833-49bd-ae85-64dbc5908057", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.620347] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.643958] env[63175]: INFO nova.scheduler.client.report [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted allocations for instance 54440032-4d9b-41d4-9ef2-5a79a4224fa6 [ 1002.867362] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52380f29-056e-035c-ce46-c7c14b42f83e, 'name': SearchDatastore_Task, 'duration_secs': 0.007994} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.867854] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.868278] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/19ec0aa6-3078-419d-849d-1bb371325a87.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.868709] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f5f0435-cf45-4901-abb7-d32082f82ff6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.877314] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1002.877314] env[63175]: value = "task-1248494" [ 1002.877314] env[63175]: _type = "Task" [ 1002.877314] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.887478] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248494, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.942404] env[63175]: DEBUG oslo_concurrency.lockutils [req-080dd458-1d04-4c65-991c-54f2f4e25971 req-7023e855-e5da-4c96-9d40-59f46493c253 service nova] Releasing lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.116391] env[63175]: DEBUG nova.compute.manager [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1003.118593] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca09f614-e638-4f88-8763-8303fb73ae7f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.155904] env[63175]: DEBUG oslo_concurrency.lockutils [None req-adb59f7e-176b-4034-8d39-66e28899dc74 tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "54440032-4d9b-41d4-9ef2-5a79a4224fa6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.625s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.313455] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "a460926e-9637-40aa-bb30-e3890a441e03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.313455] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.313966] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "a460926e-9637-40aa-bb30-e3890a441e03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.314192] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.314400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.316927] env[63175]: INFO nova.compute.manager [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Terminating instance [ 1003.388140] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248494, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.636543] env[63175]: INFO nova.compute.manager [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] instance snapshotting [ 1003.640191] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620c157c-50bb-4639-b502-aa8805aca7e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.659553] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e2de0e-b655-4bb2-9ffd-cf083c0f21c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.821169] env[63175]: DEBUG nova.compute.manager [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1003.821285] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.822084] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2929ac-41c1-4a2a-a58f-9d82a6ca9399 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.830683] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.830941] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bcb16e3-aa9c-44b5-8ba1-948ed8238ffa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.839024] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 1003.839024] env[63175]: value = "task-1248495" [ 1003.839024] env[63175]: _type = "Task" [ 1003.839024] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.847620] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.888851] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248494, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656968} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.889143] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/19ec0aa6-3078-419d-849d-1bb371325a87.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.889396] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.890225] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ff1506f-1d35-4629-a489-69e86bb805e1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.896632] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1003.896632] env[63175]: value = "task-1248496" [ 1003.896632] env[63175]: _type = "Task" [ 1003.896632] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.905065] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.174248] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1004.174579] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-450852a7-4971-43cd-b2ce-004f3f45fe07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.185258] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1004.185258] env[63175]: value = "task-1248497" [ 1004.185258] env[63175]: _type = "Task" [ 1004.185258] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.196849] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248497, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.347823] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248495, 'name': PowerOffVM_Task, 'duration_secs': 0.28597} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.347823] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.348023] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.348737] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2dc44a0c-2d65-4c3b-91a1-144d8369e458 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.406333] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078988} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.406690] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.407545] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3b3120-1d28-46d1-9ca8-6e104934f435 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.430738] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/19ec0aa6-3078-419d-849d-1bb371325a87.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.432053] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02f3bebe-7dcc-4dbf-ab5a-55791c90be73 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.446199] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.446306] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.446445] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleting the datastore file [datastore1] a460926e-9637-40aa-bb30-e3890a441e03 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.447032] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30042ab0-c40e-4082-b576-30aff815e86a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.453868] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for the task: (returnval){ [ 1004.453868] env[63175]: value = "task-1248499" [ 1004.453868] env[63175]: _type = "Task" [ 1004.453868] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.454321] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1004.454321] env[63175]: value = "task-1248500" [ 1004.454321] env[63175]: _type = "Task" [ 1004.454321] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.465254] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.468447] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248500, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.695514] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248497, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.920951] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1004.921575] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269087', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'name': 'volume-a02a3124-f4cb-440b-bde1-d22a3407142d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5', 'attached_at': '', 'detached_at': '', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'serial': 'a02a3124-f4cb-440b-bde1-d22a3407142d'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1004.923083] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca4a4d2-c6bd-46ba-905f-9a60ecede3ed {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.947116] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5b8c49-f65e-4da9-8e07-74812ff1fa52 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.973352] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-a02a3124-f4cb-440b-bde1-d22a3407142d/volume-a02a3124-f4cb-440b-bde1-d22a3407142d.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.979348] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae55d24c-e933-4091-be94-c9bf728bc440 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.000889] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248500, 'name': ReconfigVM_Task, 'duration_secs': 0.329053} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.001158] env[63175]: DEBUG oslo_vmware.api [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Task: {'id': task-1248499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264983} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.002280] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/19ec0aa6-3078-419d-849d-1bb371325a87.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.002944] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.003149] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1005.003334] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1005.003518] env[63175]: INFO nova.compute.manager [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1005.003758] env[63175]: DEBUG oslo.service.loopingcall [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.004088] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1005.004088] env[63175]: value = "task-1248501" [ 1005.004088] env[63175]: _type = "Task" [ 1005.004088] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.004222] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffc57df5-7e49-4ff2-b0fd-3925ade88c8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.005712] env[63175]: DEBUG nova.compute.manager [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1005.005827] env[63175]: DEBUG nova.network.neutron [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1005.015396] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.016613] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1005.016613] env[63175]: value = "task-1248502" [ 1005.016613] env[63175]: _type = "Task" [ 1005.016613] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.026204] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248502, 'name': Rename_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.195603] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248497, 'name': CreateSnapshot_Task, 'duration_secs': 0.787405} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.195892] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1005.196657] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7959519-a33a-459b-8147-89861f466b1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.266181] env[63175]: DEBUG nova.compute.manager [req-82f5a08a-ad68-46d2-86da-2a1d030cfe35 req-a23d8cdf-52b1-406a-8d2d-21a502db582a service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Received event network-vif-deleted-25225abe-2dcf-4815-85da-5fbf5ea4f2b5 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1005.266181] env[63175]: INFO nova.compute.manager [req-82f5a08a-ad68-46d2-86da-2a1d030cfe35 req-a23d8cdf-52b1-406a-8d2d-21a502db582a service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Neutron deleted interface 25225abe-2dcf-4815-85da-5fbf5ea4f2b5; detaching it from the instance and deleting it from the info cache [ 1005.266283] env[63175]: DEBUG nova.network.neutron [req-82f5a08a-ad68-46d2-86da-2a1d030cfe35 req-a23d8cdf-52b1-406a-8d2d-21a502db582a service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.330582] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.330912] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.518264] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.525970] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248502, 'name': Rename_Task, 'duration_secs': 0.140037} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.526296] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.526591] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-517d5410-4411-4cd2-a30f-79a6eed67841 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.532994] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1005.532994] env[63175]: value = "task-1248503" [ 1005.532994] env[63175]: _type = "Task" [ 1005.532994] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.546384] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248503, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.718968] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1005.719319] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-316bee24-7b6d-4065-86c4-81c6e1befe94 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.727188] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1005.727188] env[63175]: value = "task-1248504" [ 1005.727188] env[63175]: _type = "Task" [ 1005.727188] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.735361] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248504, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.739013] env[63175]: DEBUG nova.network.neutron [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.768767] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1547aca-c3cd-49d2-90e9-8d26485f6480 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.780674] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9be28e2-cf1b-4438-82cb-751db7d64ed6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.810291] env[63175]: DEBUG nova.compute.manager [req-82f5a08a-ad68-46d2-86da-2a1d030cfe35 req-a23d8cdf-52b1-406a-8d2d-21a502db582a service nova] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Detach interface failed, port_id=25225abe-2dcf-4815-85da-5fbf5ea4f2b5, reason: Instance a460926e-9637-40aa-bb30-e3890a441e03 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1005.833619] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1006.017588] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248501, 'name': ReconfigVM_Task, 'duration_secs': 0.68106} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.017900] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-a02a3124-f4cb-440b-bde1-d22a3407142d/volume-a02a3124-f4cb-440b-bde1-d22a3407142d.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.023062] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9d1e61e-60f9-49d8-a6d9-84db5617133a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.046761] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248503, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.048621] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1006.048621] env[63175]: value = "task-1248505" [ 1006.048621] env[63175]: _type = "Task" [ 1006.048621] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.060229] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.238055] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248504, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.241601] env[63175]: INFO nova.compute.manager [-] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Took 1.24 seconds to deallocate network for instance. [ 1006.356182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.356466] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.358491] env[63175]: INFO nova.compute.claims [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.543783] env[63175]: DEBUG oslo_vmware.api [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248503, 'name': PowerOnVM_Task, 'duration_secs': 0.543998} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.544194] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.544194] env[63175]: INFO nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Took 7.47 seconds to spawn the instance on the hypervisor. [ 1006.544373] env[63175]: DEBUG nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.545289] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0144b0-f289-4233-a9b0-b26b3b5af8ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.561078] env[63175]: DEBUG oslo_vmware.api [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248505, 'name': ReconfigVM_Task, 'duration_secs': 0.281294} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.561869] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269087', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'name': 'volume-a02a3124-f4cb-440b-bde1-d22a3407142d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5', 'attached_at': '', 'detached_at': '', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'serial': 'a02a3124-f4cb-440b-bde1-d22a3407142d'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1006.738050] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248504, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.748102] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.066974] env[63175]: INFO nova.compute.manager [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Took 14.31 seconds to build instance. [ 1007.239495] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248504, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.379622] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.380128] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.472633] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783e953b-97d5-442f-865e-eeaecbbe2c83 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.480268] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8fd0c2-f0c1-47bf-903f-7a44db4ca699 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.510114] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ee333b-2c80-4783-9047-2f22b15fb98d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.517748] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaacfa13-244d-4ffd-bcea-b2f65b772909 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.530982] env[63175]: DEBUG nova.compute.provider_tree [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.569212] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b5b0cc3d-3411-47b1-b482-f7643f594684 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.820s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.599720] env[63175]: DEBUG nova.objects.instance [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.738828] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248504, 'name': CloneVM_Task, 'duration_secs': 1.931178} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.739129] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Created linked-clone VM from snapshot [ 1007.739854] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2272e6-9260-4d09-8d99-f0699f53d5d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.747273] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Uploading image 3e1d3ed6-5f10-4719-b767-6065abde5cbe {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1007.761010] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1007.761267] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-feb4aa2a-bf22-4c81-b58a-2191e0befba6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.767399] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1007.767399] env[63175]: value = "task-1248506" [ 1007.767399] env[63175]: _type = "Task" [ 1007.767399] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.774811] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248506, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.883414] env[63175]: DEBUG nova.compute.utils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.033838] env[63175]: DEBUG nova.scheduler.client.report [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1008.106996] env[63175]: DEBUG oslo_concurrency.lockutils [None req-098593d0-cf0d-4b94-a06a-22d0e224c39a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.806s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.277962] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248506, 'name': Destroy_Task, 'duration_secs': 0.302503} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.278261] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Destroyed the VM [ 1008.278502] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1008.278760] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6af2d61e-15f1-4132-8318-782b06336614 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.285219] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1008.285219] env[63175]: value = "task-1248507" [ 1008.285219] env[63175]: _type = "Task" [ 1008.285219] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.292923] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248507, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.378590] env[63175]: INFO nova.compute.manager [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Rescuing [ 1008.378879] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.379602] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.379602] env[63175]: DEBUG nova.network.neutron [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.386642] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.538963] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.182s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.539498] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1008.542098] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.794s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.542377] env[63175]: DEBUG nova.objects.instance [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lazy-loading 'resources' on Instance uuid a460926e-9637-40aa-bb30-e3890a441e03 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.795574] env[63175]: DEBUG oslo_vmware.api [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248507, 'name': RemoveSnapshot_Task, 'duration_secs': 0.486205} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.795987] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1008.838762] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.839073] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.045161] env[63175]: DEBUG nova.compute.utils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.049761] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1009.049761] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.090608] env[63175]: DEBUG nova.policy [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.098553] env[63175]: DEBUG nova.network.neutron [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updating instance_info_cache with network_info: [{"id": "1aa15971-a833-49bd-ae85-64dbc5908057", "address": "fa:16:3e:36:52:49", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa15971-a8", "ovs_interfaceid": "1aa15971-a833-49bd-ae85-64dbc5908057", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.171445] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a46640-f090-4c69-a975-87895d86b2cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.180709] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b557500e-92f2-467b-90d0-7ef261b4685f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.210643] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfe5d1b-3edb-49ff-8e8e-3129e7bad12f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.218925] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5788122b-a767-49c8-a8c9-eeaf6dcf1ebd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.233354] env[63175]: DEBUG nova.compute.provider_tree [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.302669] env[63175]: WARNING nova.compute.manager [None req-c816fa09-cba8-4afb-84b3-225c67c5003e tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Image not found during snapshot: nova.exception.ImageNotFound: Image 3e1d3ed6-5f10-4719-b767-6065abde5cbe could not be found. [ 1009.343412] env[63175]: INFO nova.compute.manager [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Detaching volume a02a3124-f4cb-440b-bde1-d22a3407142d [ 1009.380768] env[63175]: INFO nova.virt.block_device [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Attempting to driver detach volume a02a3124-f4cb-440b-bde1-d22a3407142d from mountpoint /dev/sdb [ 1009.381117] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1009.381279] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269087', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'name': 'volume-a02a3124-f4cb-440b-bde1-d22a3407142d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5', 'attached_at': '', 'detached_at': '', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'serial': 'a02a3124-f4cb-440b-bde1-d22a3407142d'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1009.382219] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e71d7bd-1a88-48df-851f-4a666e1e10b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.406560] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Successfully created port: 72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1009.409053] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8423c9-f4b6-44df-a1a9-41c4144ea4ac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.417181] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d0505a-b62e-4e0f-bf32-a85804b6c97d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.437784] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf0cae0-0619-4d58-b64e-e56f91aa9744 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.452870] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.453162] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.453410] env[63175]: INFO nova.compute.manager [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attaching volume 98c31701-3c50-436f-853c-042900346e59 to /dev/sdb [ 1009.455291] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] The volume has not been displaced from its original location: [datastore1] volume-a02a3124-f4cb-440b-bde1-d22a3407142d/volume-a02a3124-f4cb-440b-bde1-d22a3407142d.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1009.460270] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1009.463264] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f2c3a1c-2e55-4d4d-b86b-cfdf911217d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.481059] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1009.481059] env[63175]: value = "task-1248508" [ 1009.481059] env[63175]: _type = "Task" [ 1009.481059] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.489080] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.501944] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d6386f-5a3b-4328-af81-bb70b0652ac4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.510396] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d5c28-eff9-472e-8c6b-b6d3730b3c1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.524583] env[63175]: DEBUG nova.virt.block_device [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating existing volume attachment record: a47959d0-5786-4171-82f6-d7fb543bf68d {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1009.553957] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1009.604637] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "refresh_cache-19ec0aa6-3078-419d-849d-1bb371325a87" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.737537] env[63175]: DEBUG nova.scheduler.client.report [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1009.848578] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.848960] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.849136] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.849352] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.849542] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.852154] env[63175]: INFO nova.compute.manager [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Terminating instance [ 1009.991968] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248508, 'name': ReconfigVM_Task, 'duration_secs': 0.250753} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.992300] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1009.996884] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8489c5e7-3c72-4aa2-a079-39399a26b945 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.011985] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1010.011985] env[63175]: value = "task-1248510" [ 1010.011985] env[63175]: _type = "Task" [ 1010.011985] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.020149] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248510, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.242504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.264326] env[63175]: INFO nova.scheduler.client.report [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Deleted allocations for instance a460926e-9637-40aa-bb30-e3890a441e03 [ 1010.356467] env[63175]: DEBUG nova.compute.manager [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1010.356862] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.357792] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4a54f0-ae65-428d-8042-64e302c45ffa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.365513] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.365769] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5a3cbba-42d1-47e3-832d-6b01adf57107 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.371576] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1010.371576] env[63175]: value = "task-1248511" [ 1010.371576] env[63175]: _type = "Task" [ 1010.371576] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.379291] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.522036] env[63175]: DEBUG oslo_vmware.api [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248510, 'name': ReconfigVM_Task, 'duration_secs': 0.139566} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.522347] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269087', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'name': 'volume-a02a3124-f4cb-440b-bde1-d22a3407142d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5', 'attached_at': '', 'detached_at': '', 'volume_id': 'a02a3124-f4cb-440b-bde1-d22a3407142d', 'serial': 'a02a3124-f4cb-440b-bde1-d22a3407142d'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1010.567147] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1010.593590] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.593863] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.594042] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.594238] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.594391] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.594548] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.594951] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.595224] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.595581] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.595852] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.596086] env[63175]: DEBUG nova.virt.hardware [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.596940] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1d9968-e6c6-4bcc-a526-05676d00c241 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.605624] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375b6c77-e0b2-4216-b9e5-49a3a82d2b84 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.772328] env[63175]: DEBUG oslo_concurrency.lockutils [None req-157dbcd3-546b-4275-ad98-3000eb667b5d tempest-ServerRescueNegativeTestJSON-898525982 tempest-ServerRescueNegativeTestJSON-898525982-project-member] Lock "a460926e-9637-40aa-bb30-e3890a441e03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.459s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.892446] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248511, 'name': PowerOffVM_Task, 'duration_secs': 0.449024} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.892759] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.892851] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.893136] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-162e2544-02db-4894-baee-ff58d5291efc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.901249] env[63175]: DEBUG nova.compute.manager [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Received event network-vif-plugged-72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1010.901467] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] Acquiring lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.902321] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.902321] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.902321] env[63175]: DEBUG nova.compute.manager [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] No waiting events found dispatching network-vif-plugged-72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1010.902321] env[63175]: WARNING nova.compute.manager [req-6c84d89d-c353-4bdb-8747-cb30334953e1 req-01050544-7062-479b-8b54-c2dbdbfa9ce9 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Received unexpected event network-vif-plugged-72df71c2-faab-4cc4-b2c3-e730bb625453 for instance with vm_state building and task_state spawning. [ 1010.902690] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.902753] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.903079] env[63175]: DEBUG nova.objects.instance [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid 48448a68-6b7e-488b-a7f2-9f910a2f6765 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.956312] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.956601] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.956797] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleting the datastore file [datastore1] b8c2febb-7024-49de-b34c-1c8ee492b39f {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.957100] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6357c2d-8ea6-4563-906f-c4a60bcc691c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.964705] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for the task: (returnval){ [ 1010.964705] env[63175]: value = "task-1248513" [ 1010.964705] env[63175]: _type = "Task" [ 1010.964705] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.972965] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.982036] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Successfully updated port: 72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.067756] env[63175]: DEBUG nova.objects.instance [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.141229] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.141541] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd0a75a7-87ab-472e-be5d-6bf980a97abd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.148897] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1011.148897] env[63175]: value = "task-1248514" [ 1011.148897] env[63175]: _type = "Task" [ 1011.148897] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.158576] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.475430] env[63175]: DEBUG oslo_vmware.api [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Task: {'id': task-1248513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138185} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.475685] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.476101] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.476101] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.476379] env[63175]: INFO nova.compute.manager [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1011.476531] env[63175]: DEBUG oslo.service.loopingcall [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.476740] env[63175]: DEBUG nova.compute.manager [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1011.476843] env[63175]: DEBUG nova.network.neutron [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.485175] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.485353] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.485451] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.553597] env[63175]: DEBUG nova.objects.instance [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid 48448a68-6b7e-488b-a7f2-9f910a2f6765 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.658570] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248514, 'name': PowerOffVM_Task, 'duration_secs': 0.205521} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.658841] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.659638] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7041843-a9a5-4aab-8bf8-5e2dd8b88456 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.678334] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a81428-7bed-4648-b9b6-86a4c08f663b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.710794] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.715022] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ef1d012-c607-4e87-b42f-4d1861c8bd90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.718559] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1011.718559] env[63175]: value = "task-1248515" [ 1011.718559] env[63175]: _type = "Task" [ 1011.718559] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.726827] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.857431] env[63175]: DEBUG nova.compute.manager [req-ac6a6bd4-95e7-42d1-a02a-97fcc6d8b4d5 req-26f8584d-f9f6-4556-a567-7561d15ba536 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Received event network-vif-deleted-ed916caf-e13e-4179-9349-feabcfaa75b3 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1011.857645] env[63175]: INFO nova.compute.manager [req-ac6a6bd4-95e7-42d1-a02a-97fcc6d8b4d5 req-26f8584d-f9f6-4556-a567-7561d15ba536 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Neutron deleted interface ed916caf-e13e-4179-9349-feabcfaa75b3; detaching it from the instance and deleting it from the info cache [ 1011.857861] env[63175]: DEBUG nova.network.neutron [req-ac6a6bd4-95e7-42d1-a02a-97fcc6d8b4d5 req-26f8584d-f9f6-4556-a567-7561d15ba536 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.032942] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1012.055777] env[63175]: DEBUG nova.objects.base [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance<48448a68-6b7e-488b-a7f2-9f910a2f6765> lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1012.056048] env[63175]: DEBUG nova.network.neutron [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1012.075767] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b431c58e-54f0-4e18-a297-95ba09050a11 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.237s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.155630] env[63175]: DEBUG nova.policy [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1012.190646] env[63175]: DEBUG nova.network.neutron [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updating instance_info_cache with network_info: [{"id": "72df71c2-faab-4cc4-b2c3-e730bb625453", "address": "fa:16:3e:f3:ef:b6", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72df71c2-fa", "ovs_interfaceid": "72df71c2-faab-4cc4-b2c3-e730bb625453", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.229269] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1012.229801] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.229801] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.229955] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.230073] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.230334] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2e92c0d-1636-49c3-91fe-4c61f0f3da37 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.238857] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.239058] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.239758] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f10239a-f4d5-45d8-acc7-58333aa158ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.245605] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1012.245605] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524b8008-83d4-5695-b801-38e3c0bbb2cb" [ 1012.245605] env[63175]: _type = "Task" [ 1012.245605] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.253996] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524b8008-83d4-5695-b801-38e3c0bbb2cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.312151] env[63175]: DEBUG nova.network.neutron [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.360357] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-113fc9cc-17e5-4648-822c-63fc73b04928 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.369636] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace672b9-3c4e-46c6-a878-4fc13f89921e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.396514] env[63175]: DEBUG nova.compute.manager [req-ac6a6bd4-95e7-42d1-a02a-97fcc6d8b4d5 req-26f8584d-f9f6-4556-a567-7561d15ba536 service nova] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Detach interface failed, port_id=ed916caf-e13e-4179-9349-feabcfaa75b3, reason: Instance b8c2febb-7024-49de-b34c-1c8ee492b39f could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1012.693628] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.693962] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Instance network_info: |[{"id": "72df71c2-faab-4cc4-b2c3-e730bb625453", "address": "fa:16:3e:f3:ef:b6", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72df71c2-fa", "ovs_interfaceid": "72df71c2-faab-4cc4-b2c3-e730bb625453", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1012.694426] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:ef:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72df71c2-faab-4cc4-b2c3-e730bb625453', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.702390] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating folder: Project (ea5b7d92029c40fd824817c25e39ac1f). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.702710] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6be9859-b457-4562-95ba-6d84cf1e5fbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.727566] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created folder: Project (ea5b7d92029c40fd824817c25e39ac1f) in parent group-v268956. [ 1012.727766] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating folder: Instances. Parent ref: group-v269094. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.728061] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8cc9223b-d4a4-4c47-9201-62a805e15a0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.738628] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created folder: Instances in parent group-v269094. [ 1012.738982] env[63175]: DEBUG oslo.service.loopingcall [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.739282] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.739552] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd1b5a94-bc0c-4d33-ae51-e2a62fb57f98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.771418] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524b8008-83d4-5695-b801-38e3c0bbb2cb, 'name': SearchDatastore_Task, 'duration_secs': 0.008579} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.773215] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.773215] env[63175]: value = "task-1248519" [ 1012.773215] env[63175]: _type = "Task" [ 1012.773215] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.773950] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d2c6b3e-f097-49c4-8fa5-9dbaf15d8a97 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.783171] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1012.783171] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528edc7a-20e6-d4a0-245b-6d76b93667e6" [ 1012.783171] env[63175]: _type = "Task" [ 1012.783171] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.787008] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248519, 'name': CreateVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.795462] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528edc7a-20e6-d4a0-245b-6d76b93667e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.815071] env[63175]: INFO nova.compute.manager [-] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Took 1.34 seconds to deallocate network for instance. [ 1012.988719] env[63175]: DEBUG nova.compute.manager [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Received event network-changed-72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1012.988999] env[63175]: DEBUG nova.compute.manager [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Refreshing instance network info cache due to event network-changed-72df71c2-faab-4cc4-b2c3-e730bb625453. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1012.989349] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] Acquiring lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.989530] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] Acquired lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.989709] env[63175]: DEBUG nova.network.neutron [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Refreshing network info cache for port 72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.265048] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.265440] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.265549] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.265759] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.265994] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.268265] env[63175]: INFO nova.compute.manager [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Terminating instance [ 1013.286486] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248519, 'name': CreateVM_Task, 'duration_secs': 0.359911} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.286716] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.290079] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.290262] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.290589] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.291092] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed9cd5ff-691a-492a-a7b3-1ad192a01f30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.295945] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1013.295945] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522c7232-c3fe-80ee-0b57-c3da18a1f9ff" [ 1013.295945] env[63175]: _type = "Task" [ 1013.295945] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.299807] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]528edc7a-20e6-d4a0-245b-6d76b93667e6, 'name': SearchDatastore_Task, 'duration_secs': 0.013472} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.302626] env[63175]: DEBUG oslo_concurrency.lockutils [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.302910] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. {{(pid=63175) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1013.303186] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18c47b91-e3fa-433d-ac41-27d12a0e0562 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.310802] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522c7232-c3fe-80ee-0b57-c3da18a1f9ff, 'name': SearchDatastore_Task, 'duration_secs': 0.009788} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.312016] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.312267] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.312497] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.312838] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.313073] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.313409] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1013.313409] env[63175]: value = "task-1248520" [ 1013.313409] env[63175]: _type = "Task" [ 1013.313409] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.313602] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af1cc063-6e99-40a6-84c9-265413912277 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.321880] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.322170] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.322363] env[63175]: DEBUG nova.objects.instance [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lazy-loading 'resources' on Instance uuid b8c2febb-7024-49de-b34c-1c8ee492b39f {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.327208] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248520, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.328188] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.328375] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.330069] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b746143-cf5a-435c-a47a-f36437ef5967 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.335990] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1013.335990] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529ca957-f7b9-7ee8-40e7-d188e1e1166c" [ 1013.335990] env[63175]: _type = "Task" [ 1013.335990] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.344327] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529ca957-f7b9-7ee8-40e7-d188e1e1166c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.725480] env[63175]: DEBUG nova.network.neutron [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updated VIF entry in instance network info cache for port 72df71c2-faab-4cc4-b2c3-e730bb625453. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.726025] env[63175]: DEBUG nova.network.neutron [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updating instance_info_cache with network_info: [{"id": "72df71c2-faab-4cc4-b2c3-e730bb625453", "address": "fa:16:3e:f3:ef:b6", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72df71c2-fa", "ovs_interfaceid": "72df71c2-faab-4cc4-b2c3-e730bb625453", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.772186] env[63175]: DEBUG nova.compute.manager [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1013.772417] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.773325] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da91cea2-2ae1-465a-8b73-a8f72d9b5ab3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.786489] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.786756] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0dee764a-827b-4564-9d89-07b21ff1e164 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.799023] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1013.799023] env[63175]: value = "task-1248521" [ 1013.799023] env[63175]: _type = "Task" [ 1013.799023] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.808111] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.828368] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248520, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450573} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.828903] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. [ 1013.829908] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de96dd3e-6171-48f5-bdaa-ab3f1b1519df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.856919] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.862759] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfe46ee9-e1fc-4913-920e-2f28b5c89f1f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.883183] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529ca957-f7b9-7ee8-40e7-d188e1e1166c, 'name': SearchDatastore_Task, 'duration_secs': 0.008725} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.885274] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1013.885274] env[63175]: value = "task-1248522" [ 1013.885274] env[63175]: _type = "Task" [ 1013.885274] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.885604] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-445bde7a-bec7-4520-99d7-92439da6ed8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.900319] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1013.900319] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d3487-5462-6e72-b611-0fa75ff76899" [ 1013.900319] env[63175]: _type = "Task" [ 1013.900319] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.900619] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248522, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.913245] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d3487-5462-6e72-b611-0fa75ff76899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.944535] env[63175]: DEBUG nova.network.neutron [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Successfully updated port: c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.987658] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a7a0d6-5aae-4be1-afe5-c71a49fc4858 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.997709] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00441b98-fb1f-4393-9964-75b59f9586aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.033109] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e618a35-96e1-431c-9615-e14bc4b6d5b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.041551] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b8794a-ae89-4ba5-8ff0-837f21763ab5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.056150] env[63175]: DEBUG nova.compute.provider_tree [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.075120] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1014.075387] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269093', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'name': 'volume-98c31701-3c50-436f-853c-042900346e59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'serial': '98c31701-3c50-436f-853c-042900346e59'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1014.076304] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2579aff7-6dec-4950-b0b8-1e9f131d0d10 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.093652] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ece9af-e986-4757-8772-4d3c46a17a04 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.120250] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] volume-98c31701-3c50-436f-853c-042900346e59/volume-98c31701-3c50-436f-853c-042900346e59.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.120604] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec7c4bd7-c312-4255-8653-12c00f7e57b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.139774] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1014.139774] env[63175]: value = "task-1248523" [ 1014.139774] env[63175]: _type = "Task" [ 1014.139774] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.149202] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248523, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.229834] env[63175]: DEBUG oslo_concurrency.lockutils [req-1d2c4148-995e-490b-8966-e1164cbca271 req-7e0d943f-ae09-4c59-b880-a5242f7d9283 service nova] Releasing lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.307956] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248521, 'name': PowerOffVM_Task, 'duration_secs': 0.2118} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.308286] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.308466] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.308737] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c60d634-1c23-47d2-9861-b6e358210e48 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.382038] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.382493] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.382493] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleting the datastore file [datastore2] 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.382924] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eba99ca5-0dc8-4983-b286-a7f836376050 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.392886] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1014.392886] env[63175]: value = "task-1248525" [ 1014.392886] env[63175]: _type = "Task" [ 1014.392886] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.399944] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248522, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.407624] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.412953] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d3487-5462-6e72-b611-0fa75ff76899, 'name': SearchDatastore_Task, 'duration_secs': 0.021811} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.413192] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.413459] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 5d4d277e-ec2d-40ed-b16a-6d87965888e2/5d4d277e-ec2d-40ed-b16a-6d87965888e2.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.413723] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e8ca923-dcb3-40d9-b832-3bdc3642fa84 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.421693] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1014.421693] env[63175]: value = "task-1248526" [ 1014.421693] env[63175]: _type = "Task" [ 1014.421693] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.429805] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.444998] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.445131] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.445352] env[63175]: DEBUG nova.network.neutron [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.559048] env[63175]: DEBUG nova.scheduler.client.report [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1014.654390] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248523, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.906801] env[63175]: DEBUG oslo_vmware.api [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.44704} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.911974] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.912381] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.912702] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.913038] env[63175]: INFO nova.compute.manager [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1014.913431] env[63175]: DEBUG oslo.service.loopingcall [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.913814] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248522, 'name': ReconfigVM_Task, 'duration_secs': 0.58103} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.914153] env[63175]: DEBUG nova.compute.manager [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1014.914868] env[63175]: DEBUG nova.network.neutron [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.916846] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.918333] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8f79cc-e9bc-4b3c-887f-e1d842b783af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.936550] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248526, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.978998] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45598e9a-09e5-4d4b-97f0-7bddae23e7c4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.999382] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1014.999382] env[63175]: value = "task-1248527" [ 1014.999382] env[63175]: _type = "Task" [ 1014.999382] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.012490] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248527, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.019019] env[63175]: DEBUG nova.compute.manager [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1015.019318] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.019534] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.019729] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.019924] env[63175]: DEBUG nova.compute.manager [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] No waiting events found dispatching network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1015.020130] env[63175]: WARNING nova.compute.manager [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received unexpected event network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea for instance with vm_state active and task_state None. [ 1015.020307] env[63175]: DEBUG nova.compute.manager [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1015.020468] env[63175]: DEBUG nova.compute.manager [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-c9d70bc6-f524-4fda-9b54-6b7b785161ea. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1015.020654] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.027376] env[63175]: WARNING nova.network.neutron [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 1015.064283] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.104796] env[63175]: INFO nova.scheduler.client.report [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Deleted allocations for instance b8c2febb-7024-49de-b34c-1c8ee492b39f [ 1015.151226] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248523, 'name': ReconfigVM_Task, 'duration_secs': 0.945663} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.152146] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfigured VM instance instance-00000057 to attach disk [datastore1] volume-98c31701-3c50-436f-853c-042900346e59/volume-98c31701-3c50-436f-853c-042900346e59.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.157138] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9187da0d-171f-4a20-96ee-6b1a0007ebfc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.174654] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1015.174654] env[63175]: value = "task-1248528" [ 1015.174654] env[63175]: _type = "Task" [ 1015.174654] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.187142] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.363809] env[63175]: DEBUG nova.network.neutron [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "address": "fa:16:3e:ae:b6:37", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9d70bc6-f5", "ovs_interfaceid": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.433024] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76953} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.433492] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 5d4d277e-ec2d-40ed-b16a-6d87965888e2/5d4d277e-ec2d-40ed-b16a-6d87965888e2.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.433856] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.434218] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-261cc50d-9daf-45b1-a709-3fad21d5757c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.444528] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1015.444528] env[63175]: value = "task-1248529" [ 1015.444528] env[63175]: _type = "Task" [ 1015.444528] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.452729] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.512934] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248527, 'name': ReconfigVM_Task, 'duration_secs': 0.254529} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.512934] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.514026] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9dce020-3b45-4244-adb4-6464f15a3f51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.520531] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1015.520531] env[63175]: value = "task-1248530" [ 1015.520531] env[63175]: _type = "Task" [ 1015.520531] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.529556] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.615848] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f985b3d2-d673-4c62-9bcf-d9eb694405b9 tempest-ImagesTestJSON-1689127045 tempest-ImagesTestJSON-1689127045-project-member] Lock "b8c2febb-7024-49de-b34c-1c8ee492b39f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.766s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.689394] env[63175]: DEBUG oslo_vmware.api [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248528, 'name': ReconfigVM_Task, 'duration_secs': 0.164592} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.691200] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269093', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'name': 'volume-98c31701-3c50-436f-853c-042900346e59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'serial': '98c31701-3c50-436f-853c-042900346e59'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1015.869022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.869022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.869022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.869022] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.869022] env[63175]: DEBUG nova.network.neutron [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.870079] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2ace4c-5568-4a74-9869-03d247cc1c6c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.888718] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.889334] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.889667] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.890028] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.891026] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.891026] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.891026] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.891026] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.891450] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.891794] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.892167] env[63175]: DEBUG nova.virt.hardware [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.900025] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfiguring VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1015.900025] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be400eaa-2a59-4aca-8b7e-67dfb74b80ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.922456] env[63175]: DEBUG oslo_vmware.api [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1015.922456] env[63175]: value = "task-1248531" [ 1015.922456] env[63175]: _type = "Task" [ 1015.922456] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.933283] env[63175]: DEBUG oslo_vmware.api [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248531, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.955882] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084252} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.956318] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.957468] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e773ce-6e41-4300-bbec-f3be378d0c3a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.985639] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5d4d277e-ec2d-40ed-b16a-6d87965888e2/5d4d277e-ec2d-40ed-b16a-6d87965888e2.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.988286] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91c1db4a-189c-4189-8a73-6d840f94e4f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.003148] env[63175]: DEBUG nova.network.neutron [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.010589] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1016.010589] env[63175]: value = "task-1248532" [ 1016.010589] env[63175]: _type = "Task" [ 1016.010589] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.019665] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248532, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.031387] env[63175]: DEBUG oslo_vmware.api [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248530, 'name': PowerOnVM_Task, 'duration_secs': 0.443127} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.031387] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.033974] env[63175]: DEBUG nova.compute.manager [None req-615fb517-7431-413b-91da-aa804dcb3d40 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.035371] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6283679-1ee7-4c03-9b5e-8d019c76ff40 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.434021] env[63175]: DEBUG oslo_vmware.api [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.505557] env[63175]: INFO nova.compute.manager [-] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Took 1.59 seconds to deallocate network for instance. [ 1016.530263] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248532, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.605475] env[63175]: DEBUG nova.network.neutron [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port c9d70bc6-f524-4fda-9b54-6b7b785161ea. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.605935] env[63175]: DEBUG nova.network.neutron [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "address": "fa:16:3e:ae:b6:37", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9d70bc6-f5", "ovs_interfaceid": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.732252] env[63175]: DEBUG nova.objects.instance [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.936269] env[63175]: DEBUG oslo_vmware.api [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.016301] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.017028] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.017028] env[63175]: DEBUG nova.objects.instance [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'resources' on Instance uuid 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.021340] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248532, 'name': ReconfigVM_Task, 'duration_secs': 0.529441} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.021841] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5d4d277e-ec2d-40ed-b16a-6d87965888e2/5d4d277e-ec2d-40ed-b16a-6d87965888e2.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.022515] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cce27eb-68ed-4394-bc27-100fe5379eb9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.031980] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1017.031980] env[63175]: value = "task-1248533" [ 1017.031980] env[63175]: _type = "Task" [ 1017.031980] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.045983] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248533, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.110871] env[63175]: DEBUG oslo_concurrency.lockutils [req-834fadf2-433c-4de5-9f04-89b12c78d81e req-9f89e5c5-d861-430a-8a23-64461e586cec service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.175349] env[63175]: DEBUG nova.compute.manager [req-651f67ed-a121-49e9-9130-1296e4946a5b req-92227b19-5461-47c9-b200-8b61951680e8 service nova] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Received event network-vif-deleted-83181b9f-233b-471b-bb93-9c8cafed721b {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1017.241306] env[63175]: DEBUG oslo_concurrency.lockutils [None req-a750b5f1-2fc3-42f7-a38d-4f775b593145 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.788s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.435493] env[63175]: DEBUG oslo_vmware.api [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248531, 'name': ReconfigVM_Task, 'duration_secs': 1.176346} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.436171] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.436394] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfigured VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1017.541814] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248533, 'name': Rename_Task, 'duration_secs': 0.230304} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.544482] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.544996] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6b49bdc-4c7c-4c27-8a09-4ba7fba3e5f7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.552219] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1017.552219] env[63175]: value = "task-1248534" [ 1017.552219] env[63175]: _type = "Task" [ 1017.552219] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.564577] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.624292] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cb6aa6-6291-4365-91c8-ef9d5ec969b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.632689] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf74c29-175d-4372-a0ee-ad305b78f14a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.664788] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9721c059-c5f0-439b-abdb-3c783bdb8201 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.673055] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7b8e12-5d4d-4970-950a-a1d3b645c054 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.688492] env[63175]: DEBUG nova.compute.provider_tree [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.747941] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.748233] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.941557] env[63175]: DEBUG oslo_concurrency.lockutils [None req-986a4b07-88d5-4b70-af63-546f372efd13 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.039s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.062598] env[63175]: DEBUG oslo_vmware.api [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248534, 'name': PowerOnVM_Task, 'duration_secs': 0.481231} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.062920] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.063158] env[63175]: INFO nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Took 7.50 seconds to spawn the instance on the hypervisor. [ 1018.063346] env[63175]: DEBUG nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1018.064142] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c308ef-1cf0-432d-8d7a-ce0b9c5f974b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.191109] env[63175]: DEBUG nova.scheduler.client.report [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1018.204232] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.204632] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.251131] env[63175]: DEBUG nova.compute.utils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.583950] env[63175]: INFO nova.compute.manager [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Took 12.25 seconds to build instance. [ 1018.696972] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.710581] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1018.722341] env[63175]: INFO nova.scheduler.client.report [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted allocations for instance 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5 [ 1018.753702] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.086723] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b583ce34-0fa0-4698-b0cb-6460a2f92815 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.756s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.234529] env[63175]: DEBUG oslo_concurrency.lockutils [None req-fe0bacb8-bb24-452e-b92c-c2aec3bfc73a tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.968s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.256224] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.256224] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.258967] env[63175]: INFO nova.compute.claims [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.842523] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.843396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.843708] env[63175]: INFO nova.compute.manager [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attaching volume 5d69ab8e-7ab3-43bc-b718-cb52de08ef01 to /dev/sdc [ 1019.885965] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.886171] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.892160] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbfa477-c077-450a-b475-6ce75c236d86 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.905325] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f598b1d-6e42-4abf-960d-6c5a4299aa21 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.927320] env[63175]: DEBUG nova.virt.block_device [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating existing volume attachment record: 424a2b9b-25dc-4231-aff3-3b6151193d88 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1020.384653] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc0bbdf-cbfe-4a39-a6d3-5a31bc4b2200 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.393886] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.393886] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.394803] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0105fe86-6bf0-445f-9f5e-b3c85bcec7cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.400125] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9f2e98-991b-4c46-8a7a-989a89300828 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.451236] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea18305e-a96a-444f-9c26-646d792c74d1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.454962] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e47c01c-a110-42b5-b193-b6738cd2552f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.484191] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfiguring VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1020.485779] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4768784-2f74-4f10-ad07-35cdd0f749c8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.500846] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683be7a8-b5ba-433b-928e-7a2f072871db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.518320] env[63175]: DEBUG nova.compute.provider_tree [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.520854] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1020.520854] env[63175]: value = "task-1248536" [ 1020.520854] env[63175]: _type = "Task" [ 1020.520854] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.529982] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.905306] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.905651] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.978684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "41b0611d-7249-4e81-bace-30418ca7478a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.978684] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.023047] env[63175]: DEBUG nova.scheduler.client.report [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1021.035876] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.409196] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1021.480634] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1021.531689] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.532222] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1021.541594] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.830668] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.831088] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.929530] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.929809] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.931424] env[63175]: INFO nova.compute.claims [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.002146] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.037965] env[63175]: DEBUG nova.compute.utils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.039375] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.039819] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1022.040027] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1022.083381] env[63175]: DEBUG nova.policy [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba1ccac5ed624dabb1046c8be095d389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eac46649a931423598655c2f550ce7b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.330032] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Successfully created port: e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1022.333798] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1022.539876] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.540435] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1022.854107] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.039342] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.050762] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b04a67-d738-422e-b597-8bd98c30bd65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.057756] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8734c3ff-b624-4700-83d2-bbd051d64e0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.088487] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0c5e56-f1eb-4543-9b34-6ca8b417474c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.096712] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254ce6b7-4dbc-4fe0-bdc9-a2df54ba9ab6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.112179] env[63175]: DEBUG nova.compute.provider_tree [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.540387] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.549028] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1023.573151] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.573254] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.573366] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.573561] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.573715] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.573869] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.574099] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.574269] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.574443] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.574610] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.574786] env[63175]: DEBUG nova.virt.hardware [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.575648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2cf3a1-5310-4c05-a601-197b3712265e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.584402] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d30a0d-f90d-4459-8a12-c65fb6d00920 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.614741] env[63175]: DEBUG nova.scheduler.client.report [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1023.789422] env[63175]: DEBUG nova.compute.manager [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Received event network-vif-plugged-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1023.789667] env[63175]: DEBUG oslo_concurrency.lockutils [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] Acquiring lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.790486] env[63175]: DEBUG oslo_concurrency.lockutils [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.790687] env[63175]: DEBUG oslo_concurrency.lockutils [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.790870] env[63175]: DEBUG nova.compute.manager [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] No waiting events found dispatching network-vif-plugged-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1023.791062] env[63175]: WARNING nova.compute.manager [req-e8aa5749-9f99-481f-b138-9ae37c5b59f6 req-91db0019-e430-4409-b0a0-c71ad4e9fdf1 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Received unexpected event network-vif-plugged-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a for instance with vm_state building and task_state spawning. [ 1023.880373] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Successfully updated port: e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.040743] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.119284] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.189s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.119792] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1024.122974] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.121s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.124355] env[63175]: INFO nova.compute.claims [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.383613] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.383819] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.384084] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.485106] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1024.485378] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269097', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'name': 'volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'serial': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1024.486356] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f350968e-3af4-456f-9d87-745260ce0ead {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.504625] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b29eddb-60e5-4178-b813-e72b8a5242d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.533306] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01/volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.533669] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1521c353-5771-46be-955e-22741cebe7d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.554951] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.555789] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1024.555789] env[63175]: value = "task-1248538" [ 1024.555789] env[63175]: _type = "Task" [ 1024.555789] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.564074] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.625058] env[63175]: DEBUG nova.compute.utils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.626369] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1024.627686] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.667715] env[63175]: DEBUG nova.policy [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.917806] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1025.000501] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Successfully created port: 460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.043747] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.050708] env[63175]: DEBUG nova.network.neutron [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [{"id": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "address": "fa:16:3e:d4:db:26", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d8d85d-1c", "ovs_interfaceid": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.065442] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248538, 'name': ReconfigVM_Task, 'duration_secs': 0.472733} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.066321] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01/volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.071110] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e3ec6b2-f8aa-4252-92b2-2f957e5b84be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.086110] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1025.086110] env[63175]: value = "task-1248539" [ 1025.086110] env[63175]: _type = "Task" [ 1025.086110] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.094614] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.129838] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1025.257066] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a006e272-9121-4a11-8b88-c9eed4f7f500 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.264766] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a954c9-082d-4107-b75e-c35855ad3ed4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.295755] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6819375-9bfb-4be3-9574-38d6a86f968e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.303742] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9683c9ab-7d1f-4ae8-8c4b-103577f8fc99 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.317108] env[63175]: DEBUG nova.compute.provider_tree [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.545598] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.553239] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.553592] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Instance network_info: |[{"id": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "address": "fa:16:3e:d4:db:26", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d8d85d-1c", "ovs_interfaceid": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1025.554143] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:db:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0f246b87-f105-4b33-a71d-5caf8e99e074', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0d8d85d-1c53-4c97-9040-a0bff81a2f4a', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.561371] env[63175]: DEBUG oslo.service.loopingcall [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.561589] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.561812] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2599974a-c4d5-4aa6-9bb8-ddb489ed943a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.581707] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.581707] env[63175]: value = "task-1248540" [ 1025.581707] env[63175]: _type = "Task" [ 1025.581707] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.592723] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248540, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.598918] env[63175]: DEBUG oslo_vmware.api [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248539, 'name': ReconfigVM_Task, 'duration_secs': 0.139148} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.599289] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269097', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'name': 'volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'serial': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1025.816043] env[63175]: DEBUG nova.compute.manager [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Received event network-changed-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1025.816350] env[63175]: DEBUG nova.compute.manager [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Refreshing instance network info cache due to event network-changed-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1025.816490] env[63175]: DEBUG oslo_concurrency.lockutils [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] Acquiring lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.816619] env[63175]: DEBUG oslo_concurrency.lockutils [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] Acquired lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.816801] env[63175]: DEBUG nova.network.neutron [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Refreshing network info cache for port e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.820656] env[63175]: DEBUG nova.scheduler.client.report [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1026.045307] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.092523] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248540, 'name': CreateVM_Task, 'duration_secs': 0.318775} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.092781] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.093468] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.093657] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.094018] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.094305] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdba17a8-3963-4fa0-9554-0c176c04618e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.099904] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1026.099904] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ede52f-7c58-ef36-b531-73dd9706b5b0" [ 1026.099904] env[63175]: _type = "Task" [ 1026.099904] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.113521] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52ede52f-7c58-ef36-b531-73dd9706b5b0, 'name': SearchDatastore_Task, 'duration_secs': 0.01116} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.113841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.114113] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.114432] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.114642] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.114882] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.115232] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3aa327d-b344-40be-a65e-4be6aa67f52b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.125117] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.125323] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.126054] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d2b5317-c964-479c-91af-357b9b414a76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.131682] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1026.131682] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a0a745-49c1-8a46-a087-c495b72e7c1f" [ 1026.131682] env[63175]: _type = "Task" [ 1026.131682] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.141602] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1026.143407] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a0a745-49c1-8a46-a087-c495b72e7c1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.166330] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.166635] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.166810] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.167015] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.167189] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.167344] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.167561] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.167731] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.167910] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.168099] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.168285] env[63175]: DEBUG nova.virt.hardware [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.169373] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d027df9-2865-4890-983b-edb6b5fb5c94 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.177477] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d890eb4a-6c28-4607-a213-edcd0f53eac9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.324740] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.325451] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1026.328177] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.474s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.329634] env[63175]: INFO nova.compute.claims [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.485866] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Successfully updated port: 460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.515501] env[63175]: DEBUG nova.network.neutron [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updated VIF entry in instance network info cache for port e0d8d85d-1c53-4c97-9040-a0bff81a2f4a. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1026.515950] env[63175]: DEBUG nova.network.neutron [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [{"id": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "address": "fa:16:3e:d4:db:26", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d8d85d-1c", "ovs_interfaceid": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.545270] env[63175]: DEBUG oslo_vmware.api [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248536, 'name': ReconfigVM_Task, 'duration_secs': 5.806616} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.545516] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.545720] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Reconfigured VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1026.642270] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a0a745-49c1-8a46-a087-c495b72e7c1f, 'name': SearchDatastore_Task, 'duration_secs': 0.022674} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.646302] env[63175]: DEBUG nova.objects.instance [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.647604] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def8c154-6891-4732-aeb8-59f26ce69270 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.653814] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1026.653814] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bc6b21-9c8c-8085-2977-5575c9f7762c" [ 1026.653814] env[63175]: _type = "Task" [ 1026.653814] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.662995] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bc6b21-9c8c-8085-2977-5575c9f7762c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.834881] env[63175]: DEBUG nova.compute.utils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.838323] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1026.838723] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.881023] env[63175]: DEBUG nova.policy [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9122760c9ee74855b2edab52b862613e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4bb6b47581d42f586b3af144e291547', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.988446] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.988503] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.988663] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.018743] env[63175]: DEBUG oslo_concurrency.lockutils [req-c4dd0243-3834-441a-9963-15c3bf96139d req-d35719f1-7ed2-4cda-91bb-49b32de2b37e service nova] Releasing lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.142348] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Successfully created port: 7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.153291] env[63175]: DEBUG oslo_concurrency.lockutils [None req-dcf3a791-1c41-4c78-8f7f-8860e11146dc tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.310s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.165108] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bc6b21-9c8c-8085-2977-5575c9f7762c, 'name': SearchDatastore_Task, 'duration_secs': 0.010083} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.165400] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.165659] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1027.165935] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94fe27d2-14dc-42e5-9a5b-4ffa24edc6be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.177558] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1027.177558] env[63175]: value = "task-1248541" [ 1027.177558] env[63175]: _type = "Task" [ 1027.177558] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.184866] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.342368] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1027.471321] env[63175]: DEBUG oslo_concurrency.lockutils [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.471591] env[63175]: DEBUG oslo_concurrency.lockutils [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.498256] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1639d7ec-c9db-42ce-8c4d-9eaeacacf4e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.507905] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e590c10-f1e5-460f-b14d-fe3b5159f06e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.544332] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.547047] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ded641-56e0-44eb-a75b-1ef28c7af392 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.556619] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b0fdec-81db-49df-ae47-1c5d72347da7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.577744] env[63175]: DEBUG nova.compute.provider_tree [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.685842] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248541, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476579} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.686235] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.686464] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.686730] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1772278-3a48-4894-97cf-5f143fd2b555 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.694127] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1027.694127] env[63175]: value = "task-1248542" [ 1027.694127] env[63175]: _type = "Task" [ 1027.694127] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.701971] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.710055] env[63175]: DEBUG nova.network.neutron [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Updating instance_info_cache with network_info: [{"id": "460ecde6-43d3-42d7-8ee4-882e34414c45", "address": "fa:16:3e:d0:81:b2", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460ecde6-43", "ovs_interfaceid": "460ecde6-43d3-42d7-8ee4-882e34414c45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.866323] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.866510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.866694] env[63175]: DEBUG nova.network.neutron [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.909800] env[63175]: DEBUG nova.compute.manager [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Received event network-vif-plugged-460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1027.909800] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Acquiring lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.909996] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.910214] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.910390] env[63175]: DEBUG nova.compute.manager [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] No waiting events found dispatching network-vif-plugged-460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1027.910594] env[63175]: WARNING nova.compute.manager [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Received unexpected event network-vif-plugged-460ecde6-43d3-42d7-8ee4-882e34414c45 for instance with vm_state building and task_state spawning. [ 1027.910787] env[63175]: DEBUG nova.compute.manager [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Received event network-changed-460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1027.910986] env[63175]: DEBUG nova.compute.manager [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Refreshing instance network info cache due to event network-changed-460ecde6-43d3-42d7-8ee4-882e34414c45. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1027.911183] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Acquiring lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.975406] env[63175]: INFO nova.compute.manager [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Detaching volume 98c31701-3c50-436f-853c-042900346e59 [ 1028.008111] env[63175]: INFO nova.virt.block_device [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attempting to driver detach volume 98c31701-3c50-436f-853c-042900346e59 from mountpoint /dev/sdb [ 1028.008382] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1028.008578] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269093', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'name': 'volume-98c31701-3c50-436f-853c-042900346e59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'serial': '98c31701-3c50-436f-853c-042900346e59'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1028.009451] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d64599-da55-4373-9025-eaebecb11b07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.034850] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c5d19d-54ac-45cc-8089-1bf762bbe79e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.042328] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee51c8d-dec8-4858-9abe-36e2a00d97c3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.077159] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60299d06-acf2-41e0-a447-fdfaa8b61d1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.080341] env[63175]: DEBUG nova.scheduler.client.report [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1028.097197] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] The volume has not been displaced from its original location: [datastore1] volume-98c31701-3c50-436f-853c-042900346e59/volume-98c31701-3c50-436f-853c-042900346e59.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1028.102453] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1028.103444] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0115f801-7369-4695-b31f-7611dbb4b2b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.122965] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1028.122965] env[63175]: value = "task-1248543" [ 1028.122965] env[63175]: _type = "Task" [ 1028.122965] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.132132] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.204415] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091636} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.204818] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.205473] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ce92e5-913d-4644-b31c-0dc219c1622c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.220186] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.220552] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Instance network_info: |[{"id": "460ecde6-43d3-42d7-8ee4-882e34414c45", "address": "fa:16:3e:d0:81:b2", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460ecde6-43", "ovs_interfaceid": "460ecde6-43d3-42d7-8ee4-882e34414c45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1028.229963] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.230308] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Acquired lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.230496] env[63175]: DEBUG nova.network.neutron [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Refreshing network info cache for port 460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.231968] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:81:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '460ecde6-43d3-42d7-8ee4-882e34414c45', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.240023] env[63175]: DEBUG oslo.service.loopingcall [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.240247] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de15e558-cb5b-4d64-8360-7bf1a4155037 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.257396] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.257955] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f842e8c1-df63-4fc3-9be6-93293f8eeb77 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.278910] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1028.278910] env[63175]: value = "task-1248544" [ 1028.278910] env[63175]: _type = "Task" [ 1028.278910] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.280131] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.280131] env[63175]: value = "task-1248545" [ 1028.280131] env[63175]: _type = "Task" [ 1028.280131] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.293596] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248545, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.293934] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248544, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.361962] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1028.388490] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.388823] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.388979] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.389098] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.389258] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.389412] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.389666] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.389788] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.389958] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.390223] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.390450] env[63175]: DEBUG nova.virt.hardware [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.391696] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc902651-371b-4829-953e-588e93ce7b71 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.399856] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99521906-7023-4c6c-aca8-05779b12dfa2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.564780] env[63175]: DEBUG nova.compute.manager [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received event network-vif-plugged-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1028.565230] env[63175]: DEBUG oslo_concurrency.lockutils [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] Acquiring lock "41b0611d-7249-4e81-bace-30418ca7478a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.565468] env[63175]: DEBUG oslo_concurrency.lockutils [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] Lock "41b0611d-7249-4e81-bace-30418ca7478a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.565678] env[63175]: DEBUG oslo_concurrency.lockutils [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] Lock "41b0611d-7249-4e81-bace-30418ca7478a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.565948] env[63175]: DEBUG nova.compute.manager [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] No waiting events found dispatching network-vif-plugged-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1028.566128] env[63175]: WARNING nova.compute.manager [req-174bbf9f-3787-453d-bf2c-6675497c644c req-48a9b3a5-b560-4340-9eef-7cd680f022e7 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received unexpected event network-vif-plugged-7e1877cb-140e-4619-ae3b-5de462079aeb for instance with vm_state building and task_state spawning. [ 1028.586642] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.587258] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1028.638598] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248543, 'name': ReconfigVM_Task, 'duration_secs': 0.27919} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.639997] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Successfully updated port: 7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.648262] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1028.662507] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76518327-c672-4fd2-b5b1-acba2397a878 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.687021] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1028.687021] env[63175]: value = "task-1248546" [ 1028.687021] env[63175]: _type = "Task" [ 1028.687021] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.704266] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248546, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.713212] env[63175]: INFO nova.network.neutron [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Port c9d70bc6-f524-4fda-9b54-6b7b785161ea from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1028.713689] env[63175]: DEBUG nova.network.neutron [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.793920] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248544, 'name': ReconfigVM_Task, 'duration_secs': 0.442398} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.797555] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfigured VM instance instance-0000005f to attach disk [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.798238] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248545, 'name': CreateVM_Task, 'duration_secs': 0.407918} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.798503] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45f28f0e-6444-4bfc-8249-ae2d0bc747b4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.800222] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.800990] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.801233] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.801657] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.802464] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88213ce4-788d-432d-8a30-9bf4950079ea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.808001] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1028.808001] env[63175]: value = "task-1248547" [ 1028.808001] env[63175]: _type = "Task" [ 1028.808001] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.809749] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1028.809749] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d02c52-1a66-ce6d-f645-c3380797dd51" [ 1028.809749] env[63175]: _type = "Task" [ 1028.809749] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.821034] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248547, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.824060] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d02c52-1a66-ce6d-f645-c3380797dd51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.899123] env[63175]: DEBUG nova.network.neutron [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Updated VIF entry in instance network info cache for port 460ecde6-43d3-42d7-8ee4-882e34414c45. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.899511] env[63175]: DEBUG nova.network.neutron [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Updating instance_info_cache with network_info: [{"id": "460ecde6-43d3-42d7-8ee4-882e34414c45", "address": "fa:16:3e:d0:81:b2", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap460ecde6-43", "ovs_interfaceid": "460ecde6-43d3-42d7-8ee4-882e34414c45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.066769] env[63175]: DEBUG nova.compute.manager [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1029.067052] env[63175]: DEBUG nova.compute.manager [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing instance network info cache due to event network-changed-aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1029.067268] env[63175]: DEBUG oslo_concurrency.lockutils [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] Acquiring lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.095745] env[63175]: DEBUG nova.compute.utils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.097196] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1029.097390] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.142735] env[63175]: DEBUG nova.policy [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '564ff9a1de5f43f4b3b05f3592bfa833', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54aee75c86d543a995a364ed78426ec2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.144812] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.144975] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.145156] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.200652] env[63175]: DEBUG oslo_vmware.api [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248546, 'name': ReconfigVM_Task, 'duration_secs': 0.162979} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.200992] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269093', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'name': 'volume-98c31701-3c50-436f-853c-042900346e59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '98c31701-3c50-436f-853c-042900346e59', 'serial': '98c31701-3c50-436f-853c-042900346e59'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1029.216813] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.222025] env[63175]: DEBUG oslo_concurrency.lockutils [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] Acquired lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.222025] env[63175]: DEBUG nova.network.neutron [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Refreshing network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.324824] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248547, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.327764] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d02c52-1a66-ce6d-f645-c3380797dd51, 'name': SearchDatastore_Task, 'duration_secs': 0.033345} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.329118] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.329118] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.329118] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.329118] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.329118] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.329118] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fabad20c-57d7-4ddc-9caa-2f7c150a5e1a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.339071] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.339270] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.339986] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36ab3d64-8efb-48fe-83fe-b790eba9cf4e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.346650] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1029.346650] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d108e2-907b-6d75-ec51-c26a3f72c75d" [ 1029.346650] env[63175]: _type = "Task" [ 1029.346650] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.355412] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d108e2-907b-6d75-ec51-c26a3f72c75d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.402157] env[63175]: DEBUG oslo_concurrency.lockutils [req-8df6bbf4-e050-4f84-b805-05a9581e6749 req-3c5fc29d-2c44-4bbe-996c-7727ed5cfbf4 service nova] Releasing lock "refresh_cache-6d2786d6-55e4-4d82-8a8c-b725449166a1" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.470993] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Successfully created port: ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.600688] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1029.680208] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.721766] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b479a501-faf9-492c-afff-30e2517e8122 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-48448a68-6b7e-488b-a7f2-9f910a2f6765-c9d70bc6-f524-4fda-9b54-6b7b785161ea" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.836s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.750357] env[63175]: DEBUG nova.objects.instance [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.792701] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.792992] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.793370] env[63175]: DEBUG nova.objects.instance [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'flavor' on Instance uuid 8a4877b2-6284-4727-b1fc-cf30bbc65e36 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.820957] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248547, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.858222] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d108e2-907b-6d75-ec51-c26a3f72c75d, 'name': SearchDatastore_Task, 'duration_secs': 0.009475} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.858945] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b2c025a-3404-4229-84e3-56fedffe41a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.866406] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1029.866406] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aaa332-711b-8fff-edef-ab22003e82c7" [ 1029.866406] env[63175]: _type = "Task" [ 1029.866406] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.875136] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aaa332-711b-8fff-edef-ab22003e82c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.980884] env[63175]: DEBUG nova.network.neutron [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [{"id": "7e1877cb-140e-4619-ae3b-5de462079aeb", "address": "fa:16:3e:04:3a:03", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1877cb-14", "ovs_interfaceid": "7e1877cb-140e-4619-ae3b-5de462079aeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.060155] env[63175]: DEBUG nova.compute.manager [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1030.060155] env[63175]: DEBUG nova.compute.manager [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing instance network info cache due to event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1030.060155] env[63175]: DEBUG oslo_concurrency.lockutils [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] Acquiring lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.091818] env[63175]: DEBUG nova.network.neutron [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updated VIF entry in instance network info cache for port aab7c934-a283-4031-bf42-478ef430f7ef. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.092215] env[63175]: DEBUG nova.network.neutron [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [{"id": "aab7c934-a283-4031-bf42-478ef430f7ef", "address": "fa:16:3e:7a:bb:85", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab7c934-a2", "ovs_interfaceid": "aab7c934-a283-4031-bf42-478ef430f7ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.321824] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248547, 'name': Rename_Task, 'duration_secs': 1.234687} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.322384] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1030.322644] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d1f8af1-1437-46c2-a215-8d9f15554059 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.329860] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1030.329860] env[63175]: value = "task-1248548" [ 1030.329860] env[63175]: _type = "Task" [ 1030.329860] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.342423] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.377371] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aaa332-711b-8fff-edef-ab22003e82c7, 'name': SearchDatastore_Task, 'duration_secs': 0.009847} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.377600] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.377888] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 6d2786d6-55e4-4d82-8a8c-b725449166a1/6d2786d6-55e4-4d82-8a8c-b725449166a1.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.378205] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-739eb0f7-ae49-4531-b837-d8f299302b5c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.385263] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1030.385263] env[63175]: value = "task-1248549" [ 1030.385263] env[63175]: _type = "Task" [ 1030.385263] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.394233] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.445666] env[63175]: DEBUG nova.objects.instance [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'pci_requests' on Instance uuid 8a4877b2-6284-4727-b1fc-cf30bbc65e36 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.487733] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.488041] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Instance network_info: |[{"id": "7e1877cb-140e-4619-ae3b-5de462079aeb", "address": "fa:16:3e:04:3a:03", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1877cb-14", "ovs_interfaceid": "7e1877cb-140e-4619-ae3b-5de462079aeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1030.488707] env[63175]: DEBUG oslo_concurrency.lockutils [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] Acquired lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.488911] env[63175]: DEBUG nova.network.neutron [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.490347] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:3a:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e1877cb-140e-4619-ae3b-5de462079aeb', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.498447] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Creating folder: Project (f4bb6b47581d42f586b3af144e291547). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.501805] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08b680e3-309e-44e2-8c10-75d459b09be8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.515065] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Created folder: Project (f4bb6b47581d42f586b3af144e291547) in parent group-v268956. [ 1030.515355] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Creating folder: Instances. Parent ref: group-v269100. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.515725] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-379dcd59-d4b9-48bd-90f9-e95b71901cba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.531374] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Created folder: Instances in parent group-v269100. [ 1030.531709] env[63175]: DEBUG oslo.service.loopingcall [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.531967] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.532239] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c2a7980-b4db-421a-9b4f-2169ec5d351f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.556165] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.556165] env[63175]: value = "task-1248552" [ 1030.556165] env[63175]: _type = "Task" [ 1030.556165] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.572328] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248552, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.595772] env[63175]: DEBUG oslo_concurrency.lockutils [req-3e1b61d6-cff8-491a-93de-7b9fcd54e81e req-4b5c4c4b-2c45-47c2-897c-51c09a4084a7 service nova] Releasing lock "refresh_cache-48448a68-6b7e-488b-a7f2-9f910a2f6765" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.611466] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1030.638421] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.638742] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.638881] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.639102] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.639238] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.639392] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.639608] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.639774] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.639948] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.640136] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.640341] env[63175]: DEBUG nova.virt.hardware [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.641246] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74956732-1a2c-4bd9-9282-eb99fe9edab7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.652362] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cda85b6-68aa-49b7-9057-47c40584e8bf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.765957] env[63175]: DEBUG oslo_concurrency.lockutils [None req-091184d0-619d-425c-b409-04a2fa2e762e tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.294s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.804047] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.804047] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.805886] env[63175]: DEBUG nova.network.neutron [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updated VIF entry in instance network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.806299] env[63175]: DEBUG nova.network.neutron [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [{"id": "7e1877cb-140e-4619-ae3b-5de462079aeb", "address": "fa:16:3e:04:3a:03", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1877cb-14", "ovs_interfaceid": "7e1877cb-140e-4619-ae3b-5de462079aeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.841223] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248548, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.896963] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477347} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.897391] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 6d2786d6-55e4-4d82-8a8c-b725449166a1/6d2786d6-55e4-4d82-8a8c-b725449166a1.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.897620] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.897880] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1787d03-9253-4c4a-9f27-fb7580a760e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.904552] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1030.904552] env[63175]: value = "task-1248553" [ 1030.904552] env[63175]: _type = "Task" [ 1030.904552] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.914326] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248553, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.948418] env[63175]: DEBUG nova.objects.base [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Object Instance<8a4877b2-6284-4727-b1fc-cf30bbc65e36> lazy-loaded attributes: flavor,pci_requests {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1030.948705] env[63175]: DEBUG nova.network.neutron [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1031.011847] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Successfully updated port: ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.024093] env[63175]: DEBUG nova.policy [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d3abab1a4cd49baa03c35951be00a9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab8c53b470fa4c6689aef6e5d011c3b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.067149] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248552, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.212621] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1031.212822] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing instance network info cache due to event network-changed-298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1031.213133] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.213287] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.213449] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.308725] env[63175]: INFO nova.compute.manager [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Detaching volume 5d69ab8e-7ab3-43bc-b718-cb52de08ef01 [ 1031.311174] env[63175]: DEBUG oslo_concurrency.lockutils [req-063b247a-c0ea-47b7-8e07-3b1c4c6a2028 req-22b873da-d83a-486f-aa26-8ab9ada89dee service nova] Releasing lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.342509] env[63175]: DEBUG oslo_vmware.api [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248548, 'name': PowerOnVM_Task, 'duration_secs': 0.775405} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.346022] env[63175]: INFO nova.virt.block_device [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Attempting to driver detach volume 5d69ab8e-7ab3-43bc-b718-cb52de08ef01 from mountpoint /dev/sdc [ 1031.346022] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1031.346022] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269097', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'name': 'volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'serial': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1031.346022] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.346022] env[63175]: INFO nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1031.346022] env[63175]: DEBUG nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.346022] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabc2968-16fb-4bf0-8208-9a9c9cc6df50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.349541] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a720e31-39dc-40a2-953c-6541bb207cda {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.377848] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10084d5-2ab5-468d-ad0e-7903b19c82a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.385982] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dd3fd3-c6e5-4099-8a06-673c7d111117 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.412822] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10d94cb-ed39-443a-ab84-9a4ca5376d12 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.435277] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] The volume has not been displaced from its original location: [datastore2] volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01/volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1031.441051] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfiguring VM instance instance-00000057 to detach disk 2002 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.441406] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248553, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084724} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.441648] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33de2ef7-44c2-4813-bb6c-6163df93a397 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.455856] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.456966] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6f427a-229a-433f-a2c8-c5168487c31d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.479918] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 6d2786d6-55e4-4d82-8a8c-b725449166a1/6d2786d6-55e4-4d82-8a8c-b725449166a1.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.481884] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02cedf70-04fb-4066-b17c-858023fee288 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.496264] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1031.496264] env[63175]: value = "task-1248554" [ 1031.496264] env[63175]: _type = "Task" [ 1031.496264] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.505230] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.506489] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1031.506489] env[63175]: value = "task-1248555" [ 1031.506489] env[63175]: _type = "Task" [ 1031.506489] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.515271] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.515440] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.515554] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.516916] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248555, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.566239] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248552, 'name': CreateVM_Task, 'duration_secs': 0.602206} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.566422] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.567083] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.567448] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.567585] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.567834] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a8ca2e9-af3b-4aae-8d1c-a35c12db9923 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.573158] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1031.573158] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cc14dc-79d1-17a7-c7a9-71ab49555a6b" [ 1031.573158] env[63175]: _type = "Task" [ 1031.573158] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.580780] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cc14dc-79d1-17a7-c7a9-71ab49555a6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.886663] env[63175]: INFO nova.compute.manager [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Took 12.67 seconds to build instance. [ 1031.964424] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updated VIF entry in instance network info cache for port 298629e3-8cd5-4c3d-89e7-c008ca345f71. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.964836] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.006439] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248554, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.016711] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248555, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.047738] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.082964] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cc14dc-79d1-17a7-c7a9-71ab49555a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.036939} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.083396] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.083642] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.083884] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.084056] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.084258] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.086492] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a32a7b4-d9d6-4b88-bc6e-e0547ffa7b50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.102755] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.102949] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.103692] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8e766cd-e415-453e-b053-13ac412e85ad {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.109810] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1032.109810] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527996a2-2979-93b9-512f-ee49e992b791" [ 1032.109810] env[63175]: _type = "Task" [ 1032.109810] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.121316] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527996a2-2979-93b9-512f-ee49e992b791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.189144] env[63175]: DEBUG nova.network.neutron [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating instance_info_cache with network_info: [{"id": "ebdfc338-744b-4ffc-8599-925b787e3901", "address": "fa:16:3e:2f:d3:38", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebdfc338-74", "ovs_interfaceid": "ebdfc338-744b-4ffc-8599-925b787e3901", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.391187] env[63175]: DEBUG oslo_concurrency.lockutils [None req-14a2751b-4b38-4eff-99d6-6f5d2e774f70 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.186s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.421052] env[63175]: INFO nova.compute.manager [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Rescuing [ 1032.421270] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.421371] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.421482] env[63175]: DEBUG nova.network.neutron [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.467467] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.467754] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Received event network-vif-plugged-ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1032.467931] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.468156] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.468325] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.468496] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] No waiting events found dispatching network-vif-plugged-ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1032.468666] env[63175]: WARNING nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Received unexpected event network-vif-plugged-ebdfc338-744b-4ffc-8599-925b787e3901 for instance with vm_state building and task_state spawning. [ 1032.468833] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Received event network-changed-ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1032.468990] env[63175]: DEBUG nova.compute.manager [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Refreshing instance network info cache due to event network-changed-ebdfc338-744b-4ffc-8599-925b787e3901. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1032.469170] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Acquiring lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.508091] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248554, 'name': ReconfigVM_Task, 'duration_secs': 0.59265} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.511797] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Reconfigured VM instance instance-00000057 to detach disk 2002 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1032.516343] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ddb5601-1e97-4382-8c93-06358cb60acd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.535092] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248555, 'name': ReconfigVM_Task, 'duration_secs': 0.758365} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.536429] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 6d2786d6-55e4-4d82-8a8c-b725449166a1/6d2786d6-55e4-4d82-8a8c-b725449166a1.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.537159] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1032.537159] env[63175]: value = "task-1248556" [ 1032.537159] env[63175]: _type = "Task" [ 1032.537159] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.537366] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c51b55ee-eb0d-4236-af39-722fd296fd0e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.548754] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248556, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.550179] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1032.550179] env[63175]: value = "task-1248557" [ 1032.550179] env[63175]: _type = "Task" [ 1032.550179] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.558742] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248557, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.621617] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]527996a2-2979-93b9-512f-ee49e992b791, 'name': SearchDatastore_Task, 'duration_secs': 0.014068} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.622542] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcac7acd-402f-434a-9871-75d5ddc5bb18 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.628570] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1032.628570] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aec290-383b-e864-7ae2-71db9161eccd" [ 1032.628570] env[63175]: _type = "Task" [ 1032.628570] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.637212] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aec290-383b-e864-7ae2-71db9161eccd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.692403] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.693110] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Instance network_info: |[{"id": "ebdfc338-744b-4ffc-8599-925b787e3901", "address": "fa:16:3e:2f:d3:38", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebdfc338-74", "ovs_interfaceid": "ebdfc338-744b-4ffc-8599-925b787e3901", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1032.693635] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Acquired lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.693635] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Refreshing network info cache for port ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.698087] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:d3:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebdfc338-744b-4ffc-8599-925b787e3901', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.708049] env[63175]: DEBUG oslo.service.loopingcall [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.711947] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.713392] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a0255fa-1c45-4097-9622-3c0fa9b16357 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.734635] env[63175]: DEBUG nova.compute.manager [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1032.737475] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.737475] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.737475] env[63175]: DEBUG oslo_concurrency.lockutils [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.737475] env[63175]: DEBUG nova.compute.manager [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] No waiting events found dispatching network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1032.737475] env[63175]: WARNING nova.compute.manager [req-4ee7e2bf-0256-4cf9-aa64-1ebf715c92c4 req-bd70fa82-cffc-493a-a1eb-a90b8c0883bc service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received unexpected event network-vif-plugged-c9d70bc6-f524-4fda-9b54-6b7b785161ea for instance with vm_state active and task_state None. [ 1032.737475] env[63175]: DEBUG nova.network.neutron [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Successfully updated port: c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.746450] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.746450] env[63175]: value = "task-1248558" [ 1032.746450] env[63175]: _type = "Task" [ 1032.746450] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.757459] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248558, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.925876] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updated VIF entry in instance network info cache for port ebdfc338-744b-4ffc-8599-925b787e3901. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.926488] env[63175]: DEBUG nova.network.neutron [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating instance_info_cache with network_info: [{"id": "ebdfc338-744b-4ffc-8599-925b787e3901", "address": "fa:16:3e:2f:d3:38", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebdfc338-74", "ovs_interfaceid": "ebdfc338-744b-4ffc-8599-925b787e3901", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.048485] env[63175]: DEBUG oslo_vmware.api [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248556, 'name': ReconfigVM_Task, 'duration_secs': 0.164452} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.051047] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269097', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'name': 'volume-5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '572a3821-7436-487a-a053-3819411de57e', 'attached_at': '', 'detached_at': '', 'volume_id': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01', 'serial': '5d69ab8e-7ab3-43bc-b718-cb52de08ef01'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1033.064792] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248557, 'name': Rename_Task, 'duration_secs': 0.151726} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.065174] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.065464] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2301353c-6134-4959-8f8e-f2ddb0f63d32 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.073144] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1033.073144] env[63175]: value = "task-1248559" [ 1033.073144] env[63175]: _type = "Task" [ 1033.073144] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.081809] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.127716] env[63175]: DEBUG nova.network.neutron [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [{"id": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "address": "fa:16:3e:d4:db:26", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d8d85d-1c", "ovs_interfaceid": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.138600] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aec290-383b-e864-7ae2-71db9161eccd, 'name': SearchDatastore_Task, 'duration_secs': 0.009199} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.139467] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.139833] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 41b0611d-7249-4e81-bace-30418ca7478a/41b0611d-7249-4e81-bace-30418ca7478a.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.140122] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64acdf24-2d75-4c2e-a6d4-234a41611cbe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.147746] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1033.147746] env[63175]: value = "task-1248560" [ 1033.147746] env[63175]: _type = "Task" [ 1033.147746] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.156130] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.239525] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.239775] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.240058] env[63175]: DEBUG nova.network.neutron [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.260800] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248558, 'name': CreateVM_Task, 'duration_secs': 0.328582} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.261047] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.261855] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.262085] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.262467] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.262765] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-820e4d42-b209-40d7-8d5f-cc45072943e8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.267775] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1033.267775] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5203a9c4-352a-b24a-9a58-0c6c060d2a13" [ 1033.267775] env[63175]: _type = "Task" [ 1033.267775] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.276063] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5203a9c4-352a-b24a-9a58-0c6c060d2a13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.431992] env[63175]: DEBUG oslo_concurrency.lockutils [req-ba678430-ae93-4bd5-bf38-be48db6a23c9 req-1497f964-187a-442f-b01a-71c840fa38b8 service nova] Releasing lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.584471] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248559, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.594549] env[63175]: DEBUG nova.objects.instance [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'flavor' on Instance uuid 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.633363] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.661386] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423726} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.665310] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 41b0611d-7249-4e81-bace-30418ca7478a/41b0611d-7249-4e81-bace-30418ca7478a.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.665640] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.666252] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b4d30dd-741a-46d1-8efd-4c80d6ba456d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.674803] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1033.674803] env[63175]: value = "task-1248561" [ 1033.674803] env[63175]: _type = "Task" [ 1033.674803] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.685296] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.778781] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5203a9c4-352a-b24a-9a58-0c6c060d2a13, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.779756] env[63175]: WARNING nova.network.neutron [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] a3a2ab1d-3632-446a-b9ed-b29795b0a69f already exists in list: networks containing: ['a3a2ab1d-3632-446a-b9ed-b29795b0a69f']. ignoring it [ 1033.781622] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.781886] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.782160] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.782326] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.782532] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.782836] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fef13e5-196f-4ddd-8a05-305e0545ea91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.792173] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.792474] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1033.793274] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5b526ca-e7ea-450c-bb7c-1ee6ee7a141e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.799621] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1033.799621] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e516f1-2068-1f64-aa88-08695d731550" [ 1033.799621] env[63175]: _type = "Task" [ 1033.799621] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.807744] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e516f1-2068-1f64-aa88-08695d731550, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.066730] env[63175]: DEBUG nova.network.neutron [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "address": "fa:16:3e:ae:b6:37", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9d70bc6-f5", "ovs_interfaceid": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.084060] env[63175]: DEBUG oslo_vmware.api [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248559, 'name': PowerOnVM_Task, 'duration_secs': 0.680901} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.084339] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.084548] env[63175]: INFO nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Took 7.94 seconds to spawn the instance on the hypervisor. [ 1034.084728] env[63175]: DEBUG nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1034.085558] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697ede39-c3c7-4212-a9df-3a1d45826615 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.184565] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074499} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.184846] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.185690] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55506b15-e4fd-423c-983f-67183eceb4cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.208109] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 41b0611d-7249-4e81-bace-30418ca7478a/41b0611d-7249-4e81-bace-30418ca7478a.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.208407] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-721d8067-d4a4-41c2-a7d0-926db3c2afd3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.228751] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1034.228751] env[63175]: value = "task-1248562" [ 1034.228751] env[63175]: _type = "Task" [ 1034.228751] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.237015] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248562, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.309654] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52e516f1-2068-1f64-aa88-08695d731550, 'name': SearchDatastore_Task, 'duration_secs': 0.011582} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.310425] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93a3059e-ffb9-4b23-8501-390e46839150 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.315407] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1034.315407] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520933ad-880e-0c6a-4896-2a6281e2707a" [ 1034.315407] env[63175]: _type = "Task" [ 1034.315407] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.323400] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520933ad-880e-0c6a-4896-2a6281e2707a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.570155] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.570951] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.571128] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.571930] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201123f0-8589-4f9f-8266-0b3133e29007 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.590722] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1034.590982] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1034.591167] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.591375] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1034.591540] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.591696] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1034.591906] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1034.592085] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1034.592261] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1034.592430] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1034.592606] env[63175]: DEBUG nova.virt.hardware [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.599021] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfiguring VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1034.599116] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26df5cae-c858-4349-8607-64143de5a6ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.618471] env[63175]: INFO nova.compute.manager [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Took 12.70 seconds to build instance. [ 1034.618911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ffdef61a-6542-4ca7-8063-291b44e2c648 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.815s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.628465] env[63175]: DEBUG oslo_vmware.api [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1034.628465] env[63175]: value = "task-1248563" [ 1034.628465] env[63175]: _type = "Task" [ 1034.628465] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.638326] env[63175]: DEBUG oslo_vmware.api [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248563, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.741793] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.827630] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]520933ad-880e-0c6a-4896-2a6281e2707a, 'name': SearchDatastore_Task, 'duration_secs': 0.020312} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.827930] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.828214] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 1579e4ad-f754-46ea-a9af-c85947489dcd/1579e4ad-f754-46ea-a9af-c85947489dcd.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.828714] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9ca4a84-9eb9-401c-afff-9192a0d80309 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.834288] env[63175]: DEBUG nova.compute.manager [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-changed-c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1034.834502] env[63175]: DEBUG nova.compute.manager [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing instance network info cache due to event network-changed-c9d70bc6-f524-4fda-9b54-6b7b785161ea. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1034.834728] env[63175]: DEBUG oslo_concurrency.lockutils [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.834868] env[63175]: DEBUG oslo_concurrency.lockutils [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.835119] env[63175]: DEBUG nova.network.neutron [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Refreshing network info cache for port c9d70bc6-f524-4fda-9b54-6b7b785161ea {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1034.837667] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1034.837667] env[63175]: value = "task-1248564" [ 1034.837667] env[63175]: _type = "Task" [ 1034.837667] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.847494] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248564, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.119468] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6a130ff8-1bcd-425a-a3b1-16179b8a96a8 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.214s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.139379] env[63175]: DEBUG oslo_vmware.api [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.173373] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.173991] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-466a548b-f478-40a6-abbc-13ffcf8e602f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.183598] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1035.183598] env[63175]: value = "task-1248565" [ 1035.183598] env[63175]: _type = "Task" [ 1035.183598] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.194424] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.239601] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248562, 'name': ReconfigVM_Task, 'duration_secs': 0.9611} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.239984] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 41b0611d-7249-4e81-bace-30418ca7478a/41b0611d-7249-4e81-bace-30418ca7478a.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.240675] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c02ef123-e614-4c72-b885-a01e4a788b1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.250090] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1035.250090] env[63175]: value = "task-1248566" [ 1035.250090] env[63175]: _type = "Task" [ 1035.250090] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.260984] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248566, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.311678] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.311937] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.312189] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "572a3821-7436-487a-a053-3819411de57e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.312390] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.312574] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.315163] env[63175]: INFO nova.compute.manager [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Terminating instance [ 1035.351701] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248564, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.571976] env[63175]: DEBUG nova.network.neutron [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updated VIF entry in instance network info cache for port c9d70bc6-f524-4fda-9b54-6b7b785161ea. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.572624] env[63175]: DEBUG nova.network.neutron [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "address": "fa:16:3e:ae:b6:37", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9d70bc6-f5", "ovs_interfaceid": "c9d70bc6-f524-4fda-9b54-6b7b785161ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.639031] env[63175]: DEBUG oslo_vmware.api [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.694927] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248565, 'name': PowerOffVM_Task, 'duration_secs': 0.435565} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.695279] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.696112] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10da4bdc-ddca-4611-b01f-9abd638378ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.716336] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d81248-f101-443d-8848-ad0443905309 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.746700] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.747205] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c7c2741-3a05-4ffe-acd6-b645a022b95c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.754941] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1035.754941] env[63175]: value = "task-1248567" [ 1035.754941] env[63175]: _type = "Task" [ 1035.754941] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.760908] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248566, 'name': Rename_Task, 'duration_secs': 0.362331} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.761491] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.761752] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55db7a37-cdee-4933-927b-805fb04f1351 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.767362] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1035.767638] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.767899] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.768070] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.768262] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.768491] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bffa44e-acf6-4c83-8671-17c94d659668 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.771643] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1035.771643] env[63175]: value = "task-1248568" [ 1035.771643] env[63175]: _type = "Task" [ 1035.771643] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.780049] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.781211] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.781422] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.782135] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9363b124-eb9e-4055-b3ef-b25910859e90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.787525] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1035.787525] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525222f1-f113-0689-38b6-04b5a0f3c731" [ 1035.787525] env[63175]: _type = "Task" [ 1035.787525] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.795316] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525222f1-f113-0689-38b6-04b5a0f3c731, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.819245] env[63175]: DEBUG nova.compute.manager [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1035.819462] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.820370] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b6ac4c-66d4-4a74-9521-bb9583b3b84c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.828035] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.828292] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-569cb0e6-ab7d-407c-9ad0-3db830a09379 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.835300] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1035.835300] env[63175]: value = "task-1248569" [ 1035.835300] env[63175]: _type = "Task" [ 1035.835300] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.843812] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.852577] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248564, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669963} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.852840] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 1579e4ad-f754-46ea-a9af-c85947489dcd/1579e4ad-f754-46ea-a9af-c85947489dcd.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.853153] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.853422] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ff93fdd-1197-4fa0-811e-bd6b8b1ee140 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.860387] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1035.860387] env[63175]: value = "task-1248570" [ 1035.860387] env[63175]: _type = "Task" [ 1035.860387] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.870373] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.076278] env[63175]: DEBUG oslo_concurrency.lockutils [req-1e2a0d06-e76a-4af9-bd2e-8fcd3e84237e req-e586d453-f5ad-4494-8cf6-56facbfb1abf service nova] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.140159] env[63175]: DEBUG oslo_vmware.api [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248563, 'name': ReconfigVM_Task, 'duration_secs': 1.198621} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.140990] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.141386] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfigured VM to attach interface {{(pid=63175) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1036.235128] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.235525] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.235840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.236145] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.236395] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.239349] env[63175]: INFO nova.compute.manager [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Terminating instance [ 1036.282391] env[63175]: DEBUG oslo_vmware.api [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248568, 'name': PowerOnVM_Task, 'duration_secs': 0.473427} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.282668] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.282874] env[63175]: INFO nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Took 7.92 seconds to spawn the instance on the hypervisor. [ 1036.283111] env[63175]: DEBUG nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1036.283882] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b4f478-2ca8-4dc0-8675-06f5a8301b50 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.302131] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525222f1-f113-0689-38b6-04b5a0f3c731, 'name': SearchDatastore_Task, 'duration_secs': 0.008545} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.302787] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f29fdc2-eaba-4f60-974f-0fe66cc89f3b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.308204] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1036.308204] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fda46a-b5ef-5307-c3a8-e2174271a793" [ 1036.308204] env[63175]: _type = "Task" [ 1036.308204] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.315530] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fda46a-b5ef-5307-c3a8-e2174271a793, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.345583] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248569, 'name': PowerOffVM_Task, 'duration_secs': 0.170394} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.345847] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.346031] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.346295] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-903a4f5a-1758-47f1-aaa0-f4bab10f2a97 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.369686] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064959} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.369945] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.370700] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb27351d-f0f5-4851-b347-a6fa3983e964 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.392619] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 1579e4ad-f754-46ea-a9af-c85947489dcd/1579e4ad-f754-46ea-a9af-c85947489dcd.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.392911] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78ca47a4-5125-471c-b6ae-d4a02f1a65e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.414734] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1036.414734] env[63175]: value = "task-1248572" [ 1036.414734] env[63175]: _type = "Task" [ 1036.414734] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.423517] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248572, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.453115] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.453379] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.453576] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleting the datastore file [datastore1] 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.453868] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82460a9f-f64a-4b97-891a-728a909c7fbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.460587] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for the task: (returnval){ [ 1036.460587] env[63175]: value = "task-1248573" [ 1036.460587] env[63175]: _type = "Task" [ 1036.460587] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.469102] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.647573] env[63175]: DEBUG oslo_concurrency.lockutils [None req-cf4cf78d-4457-4cde-8cce-c65491606a10 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.854s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.743442] env[63175]: DEBUG nova.compute.manager [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1036.743751] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.744798] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8439ec-2b5f-4f6f-ade4-e5a04855ad11 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.754219] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.754500] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bef2bc4-73f2-4857-a777-9c3c42816b15 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.762968] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1036.762968] env[63175]: value = "task-1248574" [ 1036.762968] env[63175]: _type = "Task" [ 1036.762968] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.774938] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.805464] env[63175]: INFO nova.compute.manager [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Took 14.82 seconds to build instance. [ 1036.819983] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fda46a-b5ef-5307-c3a8-e2174271a793, 'name': SearchDatastore_Task, 'duration_secs': 0.019738} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.821297] env[63175]: DEBUG oslo_concurrency.lockutils [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.821297] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. {{(pid=63175) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1036.822024] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-614c70e6-6490-4a99-abcb-55541ca56f6a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.830521] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1036.830521] env[63175]: value = "task-1248575" [ 1036.830521] env[63175]: _type = "Task" [ 1036.830521] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.840499] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.929581] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248572, 'name': ReconfigVM_Task, 'duration_secs': 0.488778} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.929917] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 1579e4ad-f754-46ea-a9af-c85947489dcd/1579e4ad-f754-46ea-a9af-c85947489dcd.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.930621] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f121981c-877b-4cc9-af3f-a64695b8783c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.938892] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1036.938892] env[63175]: value = "task-1248576" [ 1036.938892] env[63175]: _type = "Task" [ 1036.938892] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.948972] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248576, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.971584] env[63175]: DEBUG oslo_vmware.api [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Task: {'id': task-1248573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34536} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.971889] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.972103] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.972324] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.972474] env[63175]: INFO nova.compute.manager [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] [instance: 572a3821-7436-487a-a053-3819411de57e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1036.972734] env[63175]: DEBUG oslo.service.loopingcall [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.972978] env[63175]: DEBUG nova.compute.manager [-] [instance: 572a3821-7436-487a-a053-3819411de57e] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1036.973183] env[63175]: DEBUG nova.network.neutron [-] [instance: 572a3821-7436-487a-a053-3819411de57e] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.089153] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.089424] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.089631] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1037.275526] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248574, 'name': PowerOffVM_Task, 'duration_secs': 0.471864} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.275946] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.276096] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.276425] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7944023-b2b0-47a9-a84d-53a37962ffc3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.308479] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25bf4466-b51c-44bd-82bf-1b3bd111c37d tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.330s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.347844] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248575, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.383238] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.383469] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.383694] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore2] 6d2786d6-55e4-4d82-8a8c-b725449166a1 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.384042] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de389c12-7e42-41ba-b02d-b9407b17569e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.392131] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1037.392131] env[63175]: value = "task-1248578" [ 1037.392131] env[63175]: _type = "Task" [ 1037.392131] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.403765] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.451162] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248576, 'name': Rename_Task, 'duration_secs': 0.246266} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.455695] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.455695] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7da841d2-86db-4d7c-b993-3a16d7794808 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.461829] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1037.461829] env[63175]: value = "task-1248579" [ 1037.461829] env[63175]: _type = "Task" [ 1037.461829] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.506983] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.594058] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1037.594384] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.596492] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1037.597436] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.610873] env[63175]: DEBUG nova.compute.manager [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1037.612434] env[63175]: DEBUG nova.compute.manager [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing instance network info cache due to event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1037.612434] env[63175]: DEBUG oslo_concurrency.lockutils [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] Acquiring lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.612434] env[63175]: DEBUG oslo_concurrency.lockutils [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] Acquired lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.612434] env[63175]: DEBUG nova.network.neutron [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.841750] env[63175]: DEBUG nova.compute.manager [req-8ad54982-3065-4149-9955-49c18141f46b req-c0819cc0-49ee-440e-bd45-f1c45a663ee8 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Received event network-vif-deleted-9f3d24b2-7d86-4548-a817-f8587b820d00 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1037.841750] env[63175]: INFO nova.compute.manager [req-8ad54982-3065-4149-9955-49c18141f46b req-c0819cc0-49ee-440e-bd45-f1c45a663ee8 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Neutron deleted interface 9f3d24b2-7d86-4548-a817-f8587b820d00; detaching it from the instance and deleting it from the info cache [ 1037.841750] env[63175]: DEBUG nova.network.neutron [req-8ad54982-3065-4149-9955-49c18141f46b req-c0819cc0-49ee-440e-bd45-f1c45a663ee8 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.847051] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709263} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.847539] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk. [ 1037.848379] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa3d945-e36a-40c8-aa17-9216dc8297c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.879057] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.880310] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba7201f4-ed04-4f95-9d73-119c07556eeb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.911314] env[63175]: DEBUG oslo_vmware.api [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332538} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.913376] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.913614] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1037.913852] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1037.914543] env[63175]: INFO nova.compute.manager [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1037.914850] env[63175]: DEBUG oslo.service.loopingcall [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.915269] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1037.915269] env[63175]: value = "task-1248580" [ 1037.915269] env[63175]: _type = "Task" [ 1037.915269] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.915540] env[63175]: DEBUG nova.compute.manager [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1037.915685] env[63175]: DEBUG nova.network.neutron [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.931389] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248580, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.975874] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248579, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.100905] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.101179] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.101477] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.101477] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1038.102505] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9c6928-6326-4123-a1f6-c702f883a610 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.112656] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515c1198-2258-48ed-ad38-a74736d1bf6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.133757] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b6d24d-c541-4149-bf60-97066c38ab98 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.142196] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b243c2-f4dd-4d16-b8d7-36ab1f2f4c8b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.173323] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180284MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1038.173472] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.173677] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.212509] env[63175]: DEBUG nova.network.neutron [-] [instance: 572a3821-7436-487a-a053-3819411de57e] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.318681] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.319059] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.348131] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a04ab2e2-7791-4e77-935d-d3d9fb2fc078 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.359748] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0630f259-20e5-42f6-a5cc-51c18a1273b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.396362] env[63175]: DEBUG nova.compute.manager [req-8ad54982-3065-4149-9955-49c18141f46b req-c0819cc0-49ee-440e-bd45-f1c45a663ee8 service nova] [instance: 572a3821-7436-487a-a053-3819411de57e] Detach interface failed, port_id=9f3d24b2-7d86-4548-a817-f8587b820d00, reason: Instance 572a3821-7436-487a-a053-3819411de57e could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1038.430047] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.439668] env[63175]: DEBUG nova.network.neutron [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updated VIF entry in instance network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.440061] env[63175]: DEBUG nova.network.neutron [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [{"id": "7e1877cb-140e-4619-ae3b-5de462079aeb", "address": "fa:16:3e:04:3a:03", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1877cb-14", "ovs_interfaceid": "7e1877cb-140e-4619-ae3b-5de462079aeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.474032] env[63175]: DEBUG oslo_vmware.api [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248579, 'name': PowerOnVM_Task, 'duration_secs': 0.830906} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.474182] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.474418] env[63175]: INFO nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Took 7.86 seconds to spawn the instance on the hypervisor. [ 1038.474603] env[63175]: DEBUG nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1038.475493] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744decb4-8242-49c2-92e2-a627e249e58e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.715785] env[63175]: INFO nova.compute.manager [-] [instance: 572a3821-7436-487a-a053-3819411de57e] Took 1.74 seconds to deallocate network for instance. [ 1038.822209] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.822425] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.823497] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7a6b04-b1ff-4fea-b2d6-9fdf07a68059 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.841719] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b72160-ed59-41b0-b873-0dcb455584da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.870989] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfiguring VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1038.871398] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17030cbd-e65f-41b8-b79c-f6553a2df35c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.884775] env[63175]: DEBUG nova.network.neutron [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.893069] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1038.893069] env[63175]: value = "task-1248581" [ 1038.893069] env[63175]: _type = "Task" [ 1038.893069] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.902822] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.929533] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.943009] env[63175]: DEBUG oslo_concurrency.lockutils [req-03798287-daa1-452a-8192-6858cb527154 req-5b58c8db-0fb6-46ab-bae6-1a8f700e5704 service nova] Releasing lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.994822] env[63175]: INFO nova.compute.manager [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Took 16.16 seconds to build instance. [ 1039.212900] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 572a3821-7436-487a-a053-3819411de57e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213105] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 48448a68-6b7e-488b-a7f2-9f910a2f6765 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213236] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 8a4877b2-6284-4727-b1fc-cf30bbc65e36 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213359] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 19ec0aa6-3078-419d-849d-1bb371325a87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213479] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5d4d277e-ec2d-40ed-b16a-6d87965888e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213594] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213710] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 6d2786d6-55e4-4d82-8a8c-b725449166a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213823] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 41b0611d-7249-4e81-bace-30418ca7478a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.213935] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 1579e4ad-f754-46ea-a9af-c85947489dcd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.214148] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1039.214293] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1039.223245] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.341722] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99edf07b-a2e6-4d01-b9a1-6488e5fc6597 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.349684] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5e84ac-1286-43a4-bb77-b064c490b448 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.381736] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99623a76-0caa-48ac-84b8-7653557d25e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.387084] env[63175]: INFO nova.compute.manager [-] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Took 1.47 seconds to deallocate network for instance. [ 1039.392222] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331aa659-47d4-48b3-b49d-5439aab8a92c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.413421] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.417658] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.429473] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248580, 'name': ReconfigVM_Task, 'duration_secs': 1.317885} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.429746] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfigured VM instance instance-0000005f to attach disk [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5/8e9c4b2c-a2a7-4856-ba6d-ed187f088508-rescue.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.430613] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88833db4-7aa5-4845-8117-03e1c716b234 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.462816] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21ea5c8f-02e4-4b8c-9569-9569109f762e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.481535] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1039.481535] env[63175]: value = "task-1248582" [ 1039.481535] env[63175]: _type = "Task" [ 1039.481535] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.493047] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.497394] env[63175]: DEBUG oslo_concurrency.lockutils [None req-285c2468-32d7-414e-8280-ad3f119275f4 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.666s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.639300] env[63175]: DEBUG nova.compute.manager [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1039.639411] env[63175]: DEBUG nova.compute.manager [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing instance network info cache due to event network-changed-7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1039.639654] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] Acquiring lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.639776] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] Acquired lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.639954] env[63175]: DEBUG nova.network.neutron [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Refreshing network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.903677] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.911394] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.919423] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1039.991346] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248582, 'name': ReconfigVM_Task, 'duration_secs': 0.501213} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.991651] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.991930] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c483d5de-8ad3-4c75-8a65-5f2486c546db {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.999857] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1039.999857] env[63175]: value = "task-1248583" [ 1039.999857] env[63175]: _type = "Task" [ 1039.999857] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.012437] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.065568] env[63175]: DEBUG nova.compute.manager [req-78bd4e9f-eb3d-4c0b-adb6-2d4b76946132 req-e22947c4-57c3-4e78-8904-c17e5d0d2ea3 service nova] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Received event network-vif-deleted-460ecde6-43d3-42d7-8ee4-882e34414c45 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1040.409463] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.415170] env[63175]: DEBUG nova.network.neutron [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updated VIF entry in instance network info cache for port 7e1877cb-140e-4619-ae3b-5de462079aeb. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.415546] env[63175]: DEBUG nova.network.neutron [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [{"id": "7e1877cb-140e-4619-ae3b-5de462079aeb", "address": "fa:16:3e:04:3a:03", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e1877cb-14", "ovs_interfaceid": "7e1877cb-140e-4619-ae3b-5de462079aeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.425185] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1040.425369] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.252s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.425612] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.203s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.425840] env[63175]: DEBUG nova.objects.instance [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lazy-loading 'resources' on Instance uuid 572a3821-7436-487a-a053-3819411de57e {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.511446] env[63175]: DEBUG oslo_vmware.api [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248583, 'name': PowerOnVM_Task, 'duration_secs': 0.49961} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.511730] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.514583] env[63175]: DEBUG nova.compute.manager [None req-efa640f8-3d06-44dc-ac49-02c3bc7e1220 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1040.515467] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f2a639-6c5b-4951-b3d6-b3ccbafbd126 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.909501] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.918201] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6232c39-6e97-4552-a325-b47953b807ad req-a669a18b-143e-4faf-aa8e-37ecc6c6c09e service nova] Releasing lock "refresh_cache-41b0611d-7249-4e81-bace-30418ca7478a" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.062730] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b34500-1014-4fc7-8ad2-b00b707c5200 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.071909] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18a0d86-e131-4554-9d1b-c65e92c12a43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.103318] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c171a1c-564c-4b30-9ec6-7295d0171f88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.111426] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e6b58a-bad3-4f0c-9112-bc4b6ba8f588 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.125097] env[63175]: DEBUG nova.compute.provider_tree [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.240016] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.240296] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.410200] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.627760] env[63175]: DEBUG nova.scheduler.client.report [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1041.702757] env[63175]: INFO nova.compute.manager [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Unrescuing [ 1041.703053] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.703220] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquired lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.703396] env[63175]: DEBUG nova.network.neutron [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.742777] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1041.910921] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.090358] env[63175]: DEBUG nova.compute.manager [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Received event network-changed-ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1042.090563] env[63175]: DEBUG nova.compute.manager [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Refreshing instance network info cache due to event network-changed-ebdfc338-744b-4ffc-8599-925b787e3901. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1042.090781] env[63175]: DEBUG oslo_concurrency.lockutils [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] Acquiring lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.090931] env[63175]: DEBUG oslo_concurrency.lockutils [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] Acquired lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.091253] env[63175]: DEBUG nova.network.neutron [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Refreshing network info cache for port ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.132671] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.707s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.135147] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.231s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.135399] env[63175]: DEBUG nova.objects.instance [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 6d2786d6-55e4-4d82-8a8c-b725449166a1 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.157968] env[63175]: INFO nova.scheduler.client.report [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Deleted allocations for instance 572a3821-7436-487a-a053-3819411de57e [ 1042.261449] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.411133] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.428287] env[63175]: DEBUG nova.network.neutron [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [{"id": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "address": "fa:16:3e:d4:db:26", "network": {"id": "709fc754-b82e-491e-9870-bf8f7525b1a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1864294277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "eac46649a931423598655c2f550ce7b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0f246b87-f105-4b33-a71d-5caf8e99e074", "external-id": "nsx-vlan-transportzone-583", "segmentation_id": 583, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d8d85d-1c", "ovs_interfaceid": "e0d8d85d-1c53-4c97-9040-a0bff81a2f4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.668741] env[63175]: DEBUG oslo_concurrency.lockutils [None req-b36e69c0-ba67-4d97-9d35-da002ae3dfc9 tempest-AttachVolumeTestJSON-883245419 tempest-AttachVolumeTestJSON-883245419-project-member] Lock "572a3821-7436-487a-a053-3819411de57e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.357s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.764429] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e098583-6f85-4c09-98a0-4d41b66dbfa8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.772068] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2375c55e-27ca-4f3d-972c-5d43deda0101 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.809960] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f25611f-a7c6-4e34-8612-b19ee0a9787d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.818175] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8241ab86-b06f-4013-b110-b01481c0ec45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.832087] env[63175]: DEBUG nova.compute.provider_tree [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.846353] env[63175]: DEBUG nova.network.neutron [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updated VIF entry in instance network info cache for port ebdfc338-744b-4ffc-8599-925b787e3901. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1042.846718] env[63175]: DEBUG nova.network.neutron [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating instance_info_cache with network_info: [{"id": "ebdfc338-744b-4ffc-8599-925b787e3901", "address": "fa:16:3e:2f:d3:38", "network": {"id": "5664ea30-f55e-44b5-b7d2-21e60f8f33d3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1909607458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54aee75c86d543a995a364ed78426ec2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebdfc338-74", "ovs_interfaceid": "ebdfc338-744b-4ffc-8599-925b787e3901", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.913064] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.933333] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Releasing lock "refresh_cache-c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.934149] env[63175]: DEBUG nova.objects.instance [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lazy-loading 'flavor' on Instance uuid c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.336044] env[63175]: DEBUG nova.scheduler.client.report [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1043.348745] env[63175]: DEBUG oslo_concurrency.lockutils [req-d30612a5-6722-4763-b517-5cbe44d301f4 req-0a83d9cd-113a-423d-b61b-dcd89e07eeff service nova] Releasing lock "refresh_cache-1579e4ad-f754-46ea-a9af-c85947489dcd" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.413618] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.439334] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cfe733-d0d8-4f3a-873c-edc01600d58d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.460224] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.460518] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b99179f-3069-4348-aa19-cf765b7a0828 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.467494] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1043.467494] env[63175]: value = "task-1248585" [ 1043.467494] env[63175]: _type = "Task" [ 1043.467494] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.478525] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.841278] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.843605] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.582s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.845195] env[63175]: INFO nova.compute.claims [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.862370] env[63175]: INFO nova.scheduler.client.report [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 6d2786d6-55e4-4d82-8a8c-b725449166a1 [ 1043.913963] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.977949] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248585, 'name': PowerOffVM_Task, 'duration_secs': 0.362942} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.978250] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.983373] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1043.983649] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93af0907-9cca-49f6-9747-f7ac7f97a842 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.002600] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1044.002600] env[63175]: value = "task-1248586" [ 1044.002600] env[63175]: _type = "Task" [ 1044.002600] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.011102] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248586, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.369581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-985283fe-681a-4299-9f21-f867da970dfd tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "6d2786d6-55e4-4d82-8a8c-b725449166a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.134s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.414638] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.512302] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248586, 'name': ReconfigVM_Task, 'duration_secs': 0.367181} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.512596] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1044.512792] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1044.513093] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e06654e-426d-479c-b498-e215649e6c74 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.522010] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1044.522010] env[63175]: value = "task-1248587" [ 1044.522010] env[63175]: _type = "Task" [ 1044.522010] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.530600] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248587, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.915198] env[63175]: DEBUG oslo_vmware.api [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248581, 'name': ReconfigVM_Task, 'duration_secs': 5.773078} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.918155] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.918413] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Reconfigured VM to detach interface {{(pid=63175) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1044.984032] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634b4dd8-e30c-4222-919b-a9b7e85939c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.991831] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeb62af-d16f-4722-946f-93f6713cc419 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.030806] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ee9460-3905-427c-841f-ec37a65210ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.038703] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac93e326-7076-441a-9357-dc57a57d4b4c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.042553] env[63175]: DEBUG oslo_vmware.api [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248587, 'name': PowerOnVM_Task, 'duration_secs': 0.395893} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.042858] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.043127] env[63175]: DEBUG nova.compute.manager [None req-4ebb7b08-c94c-4a59-9e83-76ae072aeb3a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1045.044216] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff73c60-1679-4f33-9f3c-e73a7411e644 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.055302] env[63175]: DEBUG nova.compute.provider_tree [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.558407] env[63175]: DEBUG nova.scheduler.client.report [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1045.587174] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.587431] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.065055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.065738] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1046.091150] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1046.288155] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.288155] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquired lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.288155] env[63175]: DEBUG nova.network.neutron [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.418057] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.418318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.418582] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.418815] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.419019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.421298] env[63175]: INFO nova.compute.manager [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Terminating instance [ 1046.575722] env[63175]: DEBUG nova.compute.utils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.577228] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1046.577818] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.611318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.611581] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.613099] env[63175]: INFO nova.compute.claims [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.624786] env[63175]: DEBUG nova.policy [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9122760c9ee74855b2edab52b862613e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4bb6b47581d42f586b3af144e291547', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.753650] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.753923] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.754166] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.754369] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.754552] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.757160] env[63175]: INFO nova.compute.manager [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Terminating instance [ 1046.867026] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Successfully created port: 4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.924903] env[63175]: DEBUG nova.compute.manager [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1046.925337] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1046.926728] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff119088-cd6a-42fa-b110-75bef74af631 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.938405] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.938814] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-512add00-9ca5-41af-88b8-caecb9f69168 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.947542] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1046.947542] env[63175]: value = "task-1248589" [ 1046.947542] env[63175]: _type = "Task" [ 1046.947542] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.958821] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.015345] env[63175]: INFO nova.network.neutron [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Port c9d70bc6-f524-4fda-9b54-6b7b785161ea from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1047.015726] env[63175]: DEBUG nova.network.neutron [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [{"id": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "address": "fa:16:3e:68:de:fb", "network": {"id": "a3a2ab1d-3632-446a-b9ed-b29795b0a69f", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1389850306-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab8c53b470fa4c6689aef6e5d011c3b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298629e3-8c", "ovs_interfaceid": "298629e3-8cd5-4c3d-89e7-c008ca345f71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.081189] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1047.260890] env[63175]: DEBUG nova.compute.manager [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1047.261131] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.262014] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13dcadaa-a959-4dd3-a2b7-54305701b990 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.271427] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.271674] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e1929dc-7c32-41d3-b7d4-f3cea9cc497e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.278825] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1047.278825] env[63175]: value = "task-1248590" [ 1047.278825] env[63175]: _type = "Task" [ 1047.278825] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.286884] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.457501] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248589, 'name': PowerOffVM_Task, 'duration_secs': 0.23179} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.457713] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.457864] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1047.458185] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a54d2db3-38d7-4de0-8245-6fccbd1ad01e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.518720] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Releasing lock "refresh_cache-8a4877b2-6284-4727-b1fc-cf30bbc65e36" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.523237] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1047.523513] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1047.523706] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleting the datastore file [datastore2] c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.523968] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05efe32b-552d-41c6-9d90-484261ec2bfe {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.531072] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1047.531072] env[63175]: value = "task-1248592" [ 1047.531072] env[63175]: _type = "Task" [ 1047.531072] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.540271] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.742633] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e51a14-18aa-4e7a-9dbd-fdc6349f74fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.749430] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1ad3ce-390e-498a-a6b1-3407b46bb9f4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.778920] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11898420-a922-4cfe-bc5b-b0aa08933de9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.791336] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f6ede8-1d9a-4454-a221-4bb8b06db2a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.794828] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248590, 'name': PowerOffVM_Task, 'duration_secs': 0.155752} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.795098] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.795307] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1047.795810] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fc2f02b-ef65-4c57-a2ef-959240a22dc8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.806354] env[63175]: DEBUG nova.compute.provider_tree [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.868213] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1047.868521] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1047.868636] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleting the datastore file [datastore2] 8a4877b2-6284-4727-b1fc-cf30bbc65e36 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.868922] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12563143-b6dc-42c7-a376-4547e6b20c83 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.875762] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1047.875762] env[63175]: value = "task-1248594" [ 1047.875762] env[63175]: _type = "Task" [ 1047.875762] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.884560] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.024917] env[63175]: DEBUG oslo_concurrency.lockutils [None req-689850af-b0d1-47f7-bbc7-987a6141a5c8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "interface-8a4877b2-6284-4727-b1fc-cf30bbc65e36-c9d70bc6-f524-4fda-9b54-6b7b785161ea" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.706s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.041580] env[63175]: DEBUG oslo_vmware.api [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172503} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.042453] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.042699] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1048.042923] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.043168] env[63175]: INFO nova.compute.manager [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1048.043427] env[63175]: DEBUG oslo.service.loopingcall [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.043877] env[63175]: DEBUG nova.compute.manager [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1048.043975] env[63175]: DEBUG nova.network.neutron [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.090709] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1048.115515] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.115775] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.115948] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.116166] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.116347] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.116506] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.116729] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.116934] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.117173] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.117381] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.117640] env[63175]: DEBUG nova.virt.hardware [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.119070] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226eec7b-23c5-40d1-a9e7-21ce7c8769a5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.127507] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213b874f-2bf5-449f-b6f5-f5b1625cb2e8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.309672] env[63175]: DEBUG nova.scheduler.client.report [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1048.366828] env[63175]: DEBUG nova.compute.manager [req-dc2af5d3-6f3c-4de1-8345-c9aa4b0db5af req-775beab7-158a-4abf-aa3d-d0fe1f7c53b0 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Received event network-vif-deleted-e0d8d85d-1c53-4c97-9040-a0bff81a2f4a {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1048.367039] env[63175]: INFO nova.compute.manager [req-dc2af5d3-6f3c-4de1-8345-c9aa4b0db5af req-775beab7-158a-4abf-aa3d-d0fe1f7c53b0 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Neutron deleted interface e0d8d85d-1c53-4c97-9040-a0bff81a2f4a; detaching it from the instance and deleting it from the info cache [ 1048.367227] env[63175]: DEBUG nova.network.neutron [req-dc2af5d3-6f3c-4de1-8345-c9aa4b0db5af req-775beab7-158a-4abf-aa3d-d0fe1f7c53b0 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.387083] env[63175]: DEBUG oslo_vmware.api [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165191} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.387354] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.387547] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1048.387729] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.387905] env[63175]: INFO nova.compute.manager [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1048.388164] env[63175]: DEBUG oslo.service.loopingcall [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.388369] env[63175]: DEBUG nova.compute.manager [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1048.388460] env[63175]: DEBUG nova.network.neutron [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.412765] env[63175]: DEBUG nova.compute.manager [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received event network-vif-plugged-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1048.413222] env[63175]: DEBUG oslo_concurrency.lockutils [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] Acquiring lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.413597] env[63175]: DEBUG oslo_concurrency.lockutils [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.414014] env[63175]: DEBUG oslo_concurrency.lockutils [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.414238] env[63175]: DEBUG nova.compute.manager [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] No waiting events found dispatching network-vif-plugged-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1048.414534] env[63175]: WARNING nova.compute.manager [req-d7ea5fb2-b203-4938-9b0f-ed79b8429d55 req-560b358a-c244-482b-9317-4edb813f2a6f service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received unexpected event network-vif-plugged-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 for instance with vm_state building and task_state spawning. [ 1048.510181] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Successfully updated port: 4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.792582] env[63175]: DEBUG nova.network.neutron [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.814728] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.815371] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1048.870861] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8dc2a8e-c094-4635-9ec7-3e9f070f6daa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.881831] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df96f09c-e7c5-423b-b35f-18e8ac32f1cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.910849] env[63175]: DEBUG nova.compute.manager [req-dc2af5d3-6f3c-4de1-8345-c9aa4b0db5af req-775beab7-158a-4abf-aa3d-d0fe1f7c53b0 service nova] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Detach interface failed, port_id=e0d8d85d-1c53-4c97-9040-a0bff81a2f4a, reason: Instance c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1049.011930] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.012077] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.012232] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.295460] env[63175]: INFO nova.compute.manager [-] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Took 1.25 seconds to deallocate network for instance. [ 1049.321153] env[63175]: DEBUG nova.compute.utils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1049.322498] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1049.322669] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1049.374968] env[63175]: DEBUG nova.policy [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1049.546419] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.597094] env[63175]: DEBUG nova.network.neutron [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.700335] env[63175]: DEBUG nova.network.neutron [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.714631] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Successfully created port: f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1049.802683] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.802962] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.803235] env[63175]: DEBUG nova.objects.instance [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lazy-loading 'resources' on Instance uuid c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.826741] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1050.102918] env[63175]: INFO nova.compute.manager [-] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Took 1.71 seconds to deallocate network for instance. [ 1050.202840] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.203185] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Instance network_info: |[{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1050.203612] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:28:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '085fb0ff-9285-4f1d-a008-a14da4844357', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4bae9b2c-6be7-4054-99d0-6ff15c626fa0', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.211067] env[63175]: DEBUG oslo.service.loopingcall [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.211936] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.212194] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b6723ba-e657-4ce5-bcbd-7b8b6712d08f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.232207] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.232207] env[63175]: value = "task-1248596" [ 1050.232207] env[63175]: _type = "Task" [ 1050.232207] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.242724] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248596, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.396943] env[63175]: DEBUG nova.compute.manager [req-8f2bb34b-a3b7-441f-bc49-7b4de66f9456 req-dabbadcc-529e-4a69-b9f7-20fc15b79283 service nova] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Received event network-vif-deleted-298629e3-8cd5-4c3d-89e7-c008ca345f71 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1050.438134] env[63175]: DEBUG nova.compute.manager [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1050.438347] env[63175]: DEBUG nova.compute.manager [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing instance network info cache due to event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1050.438585] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] Acquiring lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.438710] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] Acquired lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.438880] env[63175]: DEBUG nova.network.neutron [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.516467] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa439db-91ab-4a6e-990b-34aa1d2694f8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.524071] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec8c429-6f8d-4a10-ad4d-706db1770f9b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.554508] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4434b9a2-f8e9-4ae0-a372-23ec3571571d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.564056] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e5db49-36ba-42c6-95de-6be7e5d1d4b8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.577811] env[63175]: DEBUG nova.compute.provider_tree [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.609339] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.742511] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248596, 'name': CreateVM_Task, 'duration_secs': 0.342723} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.742668] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.743386] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.743566] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.743879] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.744189] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d32eb63-3e2f-426a-b1ef-1155089ffc17 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.748677] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1050.748677] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218562a-4bf3-c31f-0fca-18658eb25ec9" [ 1050.748677] env[63175]: _type = "Task" [ 1050.748677] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.756247] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218562a-4bf3-c31f-0fca-18658eb25ec9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.838753] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.863017] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.863616] env[63175]: DEBUG nova.virt.hardware [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.864534] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a30899-02fe-4b09-9d17-aaeb880cfbc1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.872745] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4354b4-404e-4734-a728-fe9c65fad99f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.084019] env[63175]: DEBUG nova.scheduler.client.report [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1051.182495] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Successfully updated port: f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1051.259400] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5218562a-4bf3-c31f-0fca-18658eb25ec9, 'name': SearchDatastore_Task, 'duration_secs': 0.014582} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.259726] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.260154] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.260483] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.260656] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.260848] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.261287] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04cf297e-5292-49d8-bf37-f4b6d79057e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.264428] env[63175]: DEBUG nova.network.neutron [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updated VIF entry in instance network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1051.264891] env[63175]: DEBUG nova.network.neutron [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.270163] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.270354] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.272589] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf883050-a7b2-4ca6-87f7-4ebb0ad8cb9e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.276872] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1051.276872] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5241fc52-cfab-d58d-4771-47a659383897" [ 1051.276872] env[63175]: _type = "Task" [ 1051.276872] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.286048] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5241fc52-cfab-d58d-4771-47a659383897, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.587445] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.589833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.981s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.590234] env[63175]: DEBUG nova.objects.instance [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'resources' on Instance uuid 8a4877b2-6284-4727-b1fc-cf30bbc65e36 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.616370] env[63175]: INFO nova.scheduler.client.report [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleted allocations for instance c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5 [ 1051.690465] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.690465] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.690465] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1051.768167] env[63175]: DEBUG oslo_concurrency.lockutils [req-8b464f52-911d-4c16-8b54-aed9ae37b612 req-a69559b9-a942-4ba1-913a-173372c49898 service nova] Releasing lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.795496] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5241fc52-cfab-d58d-4771-47a659383897, 'name': SearchDatastore_Task, 'duration_secs': 0.009348} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.797068] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb484b78-5444-48d4-abaa-59277ef916ce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.804937] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1051.804937] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523169f0-20cf-729f-226d-4cb511f33c4f" [ 1051.804937] env[63175]: _type = "Task" [ 1051.804937] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.816269] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523169f0-20cf-729f-226d-4cb511f33c4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.126729] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6fdba99e-369c-4f24-b27f-8a26b84f9004 tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.708s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.220369] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a18bd5-1706-4b36-bcac-e73d3914e20b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.228505] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0621e8-5435-4e7c-ae46-5deed2314495 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.232783] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1052.265202] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197b24bf-f7f3-4c86-b5f4-b0f10a9c5541 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.273339] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc5fd4b-38f5-4bb4-851b-10cc921d4c81 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.287524] env[63175]: DEBUG nova.compute.provider_tree [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.314881] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523169f0-20cf-729f-226d-4cb511f33c4f, 'name': SearchDatastore_Task, 'duration_secs': 0.010841} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.315207] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.315463] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f99ad0a9-5a93-4033-9c57-20bc7762a48c/f99ad0a9-5a93-4033-9c57-20bc7762a48c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.315767] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de68ec36-38de-4ce2-87c0-3d81c8f5367d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.322254] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1052.322254] env[63175]: value = "task-1248597" [ 1052.322254] env[63175]: _type = "Task" [ 1052.322254] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.329554] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.431099] env[63175]: DEBUG nova.compute.manager [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Received event network-vif-plugged-f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1052.431099] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Acquiring lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.431099] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.431099] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.431099] env[63175]: DEBUG nova.compute.manager [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] No waiting events found dispatching network-vif-plugged-f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1052.431099] env[63175]: WARNING nova.compute.manager [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Received unexpected event network-vif-plugged-f0f01168-c8b5-446b-9224-a09b4e7843aa for instance with vm_state building and task_state spawning. [ 1052.431099] env[63175]: DEBUG nova.compute.manager [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Received event network-changed-f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1052.431099] env[63175]: DEBUG nova.compute.manager [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Refreshing instance network info cache due to event network-changed-f0f01168-c8b5-446b-9224-a09b4e7843aa. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1052.431099] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Acquiring lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.431879] env[63175]: DEBUG nova.network.neutron [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Updating instance_info_cache with network_info: [{"id": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "address": "fa:16:3e:41:dd:63", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0f01168-c8", "ovs_interfaceid": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.790868] env[63175]: DEBUG nova.scheduler.client.report [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1052.843674] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248597, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447204} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.843953] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] f99ad0a9-5a93-4033-9c57-20bc7762a48c/f99ad0a9-5a93-4033-9c57-20bc7762a48c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.844209] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.844472] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7c357d9-2149-4441-b034-768095e94246 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.857144] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1052.857144] env[63175]: value = "task-1248598" [ 1052.857144] env[63175]: _type = "Task" [ 1052.857144] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.867887] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.934972] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.935145] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Instance network_info: |[{"id": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "address": "fa:16:3e:41:dd:63", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0f01168-c8", "ovs_interfaceid": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1052.935567] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Acquired lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.935790] env[63175]: DEBUG nova.network.neutron [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Refreshing network info cache for port f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.937222] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:dd:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0f01168-c8b5-446b-9224-a09b4e7843aa', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1052.947346] env[63175]: DEBUG oslo.service.loopingcall [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.951379] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1052.952521] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf0d6a8d-8da6-4837-a2de-7e5c508662b2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.975670] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1052.975670] env[63175]: value = "task-1248599" [ 1052.975670] env[63175]: _type = "Task" [ 1052.975670] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.985321] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248599, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.235459] env[63175]: DEBUG nova.network.neutron [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Updated VIF entry in instance network info cache for port f0f01168-c8b5-446b-9224-a09b4e7843aa. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1053.235860] env[63175]: DEBUG nova.network.neutron [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Updating instance_info_cache with network_info: [{"id": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "address": "fa:16:3e:41:dd:63", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0f01168-c8", "ovs_interfaceid": "f0f01168-c8b5-446b-9224-a09b4e7843aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.295903] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.307662] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "19ec0aa6-3078-419d-849d-1bb371325a87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.307950] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.308278] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.308503] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.308686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.310771] env[63175]: INFO nova.compute.manager [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Terminating instance [ 1053.330125] env[63175]: INFO nova.scheduler.client.report [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted allocations for instance 8a4877b2-6284-4727-b1fc-cf30bbc65e36 [ 1053.368354] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07213} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.368589] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.369381] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371810c3-168d-4779-8925-9a6871ad5b7c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.392542] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] f99ad0a9-5a93-4033-9c57-20bc7762a48c/f99ad0a9-5a93-4033-9c57-20bc7762a48c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.393606] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7179bab7-c22c-4cc8-9171-d4679aa83c1d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.412989] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1053.412989] env[63175]: value = "task-1248600" [ 1053.412989] env[63175]: _type = "Task" [ 1053.412989] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.421371] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.486252] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248599, 'name': CreateVM_Task, 'duration_secs': 0.423848} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.486426] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1053.487137] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.487313] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.487639] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1053.487896] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cbbe45b-020b-4b19-9da1-39edb083f3d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.492685] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1053.492685] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fd149f-b22c-36f3-1e94-ae01c2858dec" [ 1053.492685] env[63175]: _type = "Task" [ 1053.492685] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.499899] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fd149f-b22c-36f3-1e94-ae01c2858dec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.738882] env[63175]: DEBUG oslo_concurrency.lockutils [req-2371e323-b7a9-498c-8add-4235165b9ae5 req-ccc3d93b-106e-4747-97a4-2d737b3e2575 service nova] Releasing lock "refresh_cache-01025311-e8d6-4fc5-9e3f-e70e3aa3302d" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.815124] env[63175]: DEBUG nova.compute.manager [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1053.815424] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.816351] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f040f82-42c5-48a6-8fc6-910b1d2410be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.823920] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.824186] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9866f97-ea04-454f-9537-d51543670118 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.829633] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1053.829633] env[63175]: value = "task-1248601" [ 1053.829633] env[63175]: _type = "Task" [ 1053.829633] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.839494] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.840122] env[63175]: DEBUG oslo_concurrency.lockutils [None req-648e4f05-2119-4a80-bff1-d9b596b97922 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "8a4877b2-6284-4727-b1fc-cf30bbc65e36" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.086s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.923991] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248600, 'name': ReconfigVM_Task, 'duration_secs': 0.286459} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.924339] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Reconfigured VM instance instance-00000063 to attach disk [datastore1] f99ad0a9-5a93-4033-9c57-20bc7762a48c/f99ad0a9-5a93-4033-9c57-20bc7762a48c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.925035] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddd0b191-76ca-48ff-a4eb-aaad030c9142 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.931926] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1053.931926] env[63175]: value = "task-1248602" [ 1053.931926] env[63175]: _type = "Task" [ 1053.931926] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.941192] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248602, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.008210] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fd149f-b22c-36f3-1e94-ae01c2858dec, 'name': SearchDatastore_Task, 'duration_secs': 0.009847} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.008718] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.009123] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.009503] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.009767] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.010110] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.010521] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca67fd95-cea8-433f-9bf0-ea52d2a4f597 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.019812] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.020047] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.020819] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05d63760-fb4f-4e03-a538-18d8e5dbd05d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.025930] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1054.025930] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522944f9-fa30-6cef-6f9f-74b30671a49e" [ 1054.025930] env[63175]: _type = "Task" [ 1054.025930] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.036137] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522944f9-fa30-6cef-6f9f-74b30671a49e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.340246] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248601, 'name': PowerOffVM_Task, 'duration_secs': 0.238378} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.340516] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.340685] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.340948] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-427c8306-8dd8-47db-8d50-a59e9778281f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.429479] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.429703] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.429891] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleting the datastore file [datastore1] 19ec0aa6-3078-419d-849d-1bb371325a87 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.430193] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccd162ed-8aa3-43de-b0be-345c5d6a353e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.436959] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for the task: (returnval){ [ 1054.436959] env[63175]: value = "task-1248604" [ 1054.436959] env[63175]: _type = "Task" [ 1054.436959] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.442588] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248602, 'name': Rename_Task, 'duration_secs': 0.182239} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.443156] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.443386] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb177330-df51-445d-9f28-63073d7d1793 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.447552] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.451643] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1054.451643] env[63175]: value = "task-1248605" [ 1054.451643] env[63175]: _type = "Task" [ 1054.451643] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.461874] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248605, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.536078] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]522944f9-fa30-6cef-6f9f-74b30671a49e, 'name': SearchDatastore_Task, 'duration_secs': 0.008097} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.536986] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db98e9e9-0cb4-4ae6-bdc3-3fcac599ae48 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.542091] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1054.542091] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c4460c-0d4a-54bc-f053-96532803bac0" [ 1054.542091] env[63175]: _type = "Task" [ 1054.542091] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.549260] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c4460c-0d4a-54bc-f053-96532803bac0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.676159] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.676489] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.676740] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.676958] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.677177] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.679475] env[63175]: INFO nova.compute.manager [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Terminating instance [ 1054.948765] env[63175]: DEBUG oslo_vmware.api [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Task: {'id': task-1248604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26639} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.949148] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.949286] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.949595] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.949881] env[63175]: INFO nova.compute.manager [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1054.950182] env[63175]: DEBUG oslo.service.loopingcall [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.950419] env[63175]: DEBUG nova.compute.manager [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1054.950523] env[63175]: DEBUG nova.network.neutron [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1054.961418] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248605, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.051785] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c4460c-0d4a-54bc-f053-96532803bac0, 'name': SearchDatastore_Task, 'duration_secs': 0.017478} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.052663] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.052663] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 01025311-e8d6-4fc5-9e3f-e70e3aa3302d/01025311-e8d6-4fc5-9e3f-e70e3aa3302d.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.052663] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c77fe34-b0ec-4e33-93a2-ab7ccbacdbf1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.058970] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1055.058970] env[63175]: value = "task-1248606" [ 1055.058970] env[63175]: _type = "Task" [ 1055.058970] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.067656] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248606, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.183560] env[63175]: DEBUG nova.compute.manager [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1055.184265] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.185318] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9901a51-ea54-4e57-90db-c9919a3594c0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.193648] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.193901] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bbaf05b-13c4-4717-9517-cf11dfb4964b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.201507] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1055.201507] env[63175]: value = "task-1248607" [ 1055.201507] env[63175]: _type = "Task" [ 1055.201507] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.207759] env[63175]: DEBUG nova.compute.manager [req-b918dde7-57dc-412c-b4c0-13ba4ed16db9 req-07af5401-c8b9-4066-9176-3b31a6932a68 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Received event network-vif-deleted-1aa15971-a833-49bd-ae85-64dbc5908057 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1055.208016] env[63175]: INFO nova.compute.manager [req-b918dde7-57dc-412c-b4c0-13ba4ed16db9 req-07af5401-c8b9-4066-9176-3b31a6932a68 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Neutron deleted interface 1aa15971-a833-49bd-ae85-64dbc5908057; detaching it from the instance and deleting it from the info cache [ 1055.208242] env[63175]: DEBUG nova.network.neutron [req-b918dde7-57dc-412c-b4c0-13ba4ed16db9 req-07af5401-c8b9-4066-9176-3b31a6932a68 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.213304] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.463983] env[63175]: DEBUG oslo_vmware.api [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248605, 'name': PowerOnVM_Task, 'duration_secs': 0.522534} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.464293] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.464521] env[63175]: INFO nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1055.464706] env[63175]: DEBUG nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1055.465568] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6182d37-8811-49ec-ae0a-628dfa941328 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.569859] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248606, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.676339] env[63175]: DEBUG nova.network.neutron [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.711262] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248607, 'name': PowerOffVM_Task, 'duration_secs': 0.206108} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.711522] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.711694] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.711939] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51eabc30-7867-45fe-aca1-d91d50f10e61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.713982] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42bac5f9-0474-4746-9f6b-6b93cdeb7008 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.722615] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ab0b95-fa62-41e5-88ef-3e0aa3b2cf18 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.749017] env[63175]: DEBUG nova.compute.manager [req-b918dde7-57dc-412c-b4c0-13ba4ed16db9 req-07af5401-c8b9-4066-9176-3b31a6932a68 service nova] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Detach interface failed, port_id=1aa15971-a833-49bd-ae85-64dbc5908057, reason: Instance 19ec0aa6-3078-419d-849d-1bb371325a87 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1055.774183] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1055.774457] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1055.774658] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleting the datastore file [datastore2] 48448a68-6b7e-488b-a7f2-9f910a2f6765 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.774919] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b27e297-b1fc-4536-b758-23fa3cbb18e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.781527] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for the task: (returnval){ [ 1055.781527] env[63175]: value = "task-1248609" [ 1055.781527] env[63175]: _type = "Task" [ 1055.781527] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.789113] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.983324] env[63175]: INFO nova.compute.manager [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Took 13.74 seconds to build instance. [ 1056.071600] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248606, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555376} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.071878] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 01025311-e8d6-4fc5-9e3f-e70e3aa3302d/01025311-e8d6-4fc5-9e3f-e70e3aa3302d.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.072112] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.072378] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58720e95-ddac-433f-bea3-5e13591e499b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.078618] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1056.078618] env[63175]: value = "task-1248610" [ 1056.078618] env[63175]: _type = "Task" [ 1056.078618] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.085800] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.178939] env[63175]: INFO nova.compute.manager [-] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Took 1.23 seconds to deallocate network for instance. [ 1056.292052] env[63175]: DEBUG oslo_vmware.api [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Task: {'id': task-1248609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141825} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.292322] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.292584] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.292821] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.293062] env[63175]: INFO nova.compute.manager [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1056.293354] env[63175]: DEBUG oslo.service.loopingcall [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.293581] env[63175]: DEBUG nova.compute.manager [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1056.293702] env[63175]: DEBUG nova.network.neutron [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1056.485264] env[63175]: DEBUG oslo_concurrency.lockutils [None req-25c5f0fc-5701-4826-9765-89dcb83ecfd8 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.245s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.590696] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067127} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.590978] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1056.591797] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6737e33c-e5e2-40ab-bf17-5f50110887fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.614908] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 01025311-e8d6-4fc5-9e3f-e70e3aa3302d/01025311-e8d6-4fc5-9e3f-e70e3aa3302d.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.615440] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74f98b9c-62b5-40a8-b3d9-2a7db88b40d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.634362] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1056.634362] env[63175]: value = "task-1248611" [ 1056.634362] env[63175]: _type = "Task" [ 1056.634362] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.642226] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.685384] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.685665] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.685899] env[63175]: DEBUG nova.objects.instance [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lazy-loading 'resources' on Instance uuid 19ec0aa6-3078-419d-849d-1bb371325a87 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.862492] env[63175]: DEBUG nova.compute.manager [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1056.862704] env[63175]: DEBUG nova.compute.manager [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing instance network info cache due to event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1056.862920] env[63175]: DEBUG oslo_concurrency.lockutils [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] Acquiring lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.863096] env[63175]: DEBUG oslo_concurrency.lockutils [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] Acquired lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.863308] env[63175]: DEBUG nova.network.neutron [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.088904] env[63175]: DEBUG nova.network.neutron [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.143412] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248611, 'name': ReconfigVM_Task, 'duration_secs': 0.452295} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.143692] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 01025311-e8d6-4fc5-9e3f-e70e3aa3302d/01025311-e8d6-4fc5-9e3f-e70e3aa3302d.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.144313] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a54e6dd6-5316-4db2-a56f-9807915670d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.150710] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1057.150710] env[63175]: value = "task-1248612" [ 1057.150710] env[63175]: _type = "Task" [ 1057.150710] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.159394] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248612, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.234136] env[63175]: DEBUG nova.compute.manager [req-604753d1-532f-4af8-8ace-60efc5957e54 req-a0f1e703-8618-4ee8-9335-6f9bd9221b82 service nova] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Received event network-vif-deleted-aab7c934-a283-4031-bf42-478ef430f7ef {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1057.288219] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b295a1c4-8c76-44d8-a534-ee84546f18a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.295745] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d162ac7e-db6e-4cfd-b8c3-b05df2bda304 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.325930] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006ae0ad-6d82-4ff1-adc3-54b9c7de5873 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.333137] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27de947-0dd5-41a7-a3f9-6f121d9d4249 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.345798] env[63175]: DEBUG nova.compute.provider_tree [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.472841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.472841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.472841] env[63175]: INFO nova.compute.manager [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Rebooting instance [ 1057.590993] env[63175]: INFO nova.compute.manager [-] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Took 1.30 seconds to deallocate network for instance. [ 1057.618284] env[63175]: DEBUG nova.network.neutron [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updated VIF entry in instance network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.618723] env[63175]: DEBUG nova.network.neutron [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.660280] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248612, 'name': Rename_Task, 'duration_secs': 0.155869} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.660571] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1057.660819] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c069cc1c-d627-4d50-a186-44390e216215 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.667492] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1057.667492] env[63175]: value = "task-1248613" [ 1057.667492] env[63175]: _type = "Task" [ 1057.667492] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.677687] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.848677] env[63175]: DEBUG nova.scheduler.client.report [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1057.996021] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.099560] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.121462] env[63175]: DEBUG oslo_concurrency.lockutils [req-b00c04b5-5f52-484a-b7e6-fcf88cbbb3f9 req-34e98b38-8ef6-471d-a7f4-77a1aa09187c service nova] Releasing lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.122268] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquired lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.122466] env[63175]: DEBUG nova.network.neutron [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.177256] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248613, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.354040] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.668s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.356494] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.257s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.356746] env[63175]: DEBUG nova.objects.instance [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lazy-loading 'resources' on Instance uuid 48448a68-6b7e-488b-a7f2-9f910a2f6765 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.377330] env[63175]: INFO nova.scheduler.client.report [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Deleted allocations for instance 19ec0aa6-3078-419d-849d-1bb371325a87 [ 1058.682231] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248613, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.850805] env[63175]: DEBUG nova.network.neutron [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.884920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4140de53-1043-40b6-868b-8799c82ae08a tempest-ServerRescueTestJSON-1856801662 tempest-ServerRescueTestJSON-1856801662-project-member] Lock "19ec0aa6-3078-419d-849d-1bb371325a87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.577s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.949685] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6dbf78-b2f1-44a9-bc21-d6e77946ac35 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.957535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c1c627-741f-4cdd-8339-21be24af08ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.987782] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef1ff37-54ab-40b7-9147-85ce39db730f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.994866] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a8656a-59ef-44a7-b679-d66e820a5e37 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.009501] env[63175]: DEBUG nova.compute.provider_tree [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.180514] env[63175]: DEBUG oslo_vmware.api [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248613, 'name': PowerOnVM_Task, 'duration_secs': 1.170229} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.180867] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.180987] env[63175]: INFO nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1059.181187] env[63175]: DEBUG nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1059.181930] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7dbe8d-ea5f-4d7e-8846-725c46f0eb76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.353502] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Releasing lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.512580] env[63175]: DEBUG nova.scheduler.client.report [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1059.699875] env[63175]: INFO nova.compute.manager [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Took 13.10 seconds to build instance. [ 1059.858418] env[63175]: DEBUG nova.compute.manager [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1059.859320] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af44d8a-2e86-47df-9368-405d3d0485bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.020104] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.662s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.049202] env[63175]: INFO nova.scheduler.client.report [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Deleted allocations for instance 48448a68-6b7e-488b-a7f2-9f910a2f6765 [ 1060.200686] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d4ab37b2-839c-4592-be17-dcf02d733a18 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.613s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.556297] env[63175]: DEBUG oslo_concurrency.lockutils [None req-47a42e8b-8b1b-4751-abef-46ef1bba6ae8 tempest-AttachInterfacesTestJSON-2135768032 tempest-AttachInterfacesTestJSON-2135768032-project-member] Lock "48448a68-6b7e-488b-a7f2-9f910a2f6765" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.880s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.568984] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.569318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.569535] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.569728] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.569898] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.572131] env[63175]: INFO nova.compute.manager [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Terminating instance [ 1060.875121] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e97960-1287-4188-9434-612c51a18ba1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.884085] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Doing hard reboot of VM {{(pid=63175) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1060.884335] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-afe7ba63-7485-495b-b530-5c160b244199 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.889940] env[63175]: DEBUG oslo_vmware.api [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1060.889940] env[63175]: value = "task-1248614" [ 1060.889940] env[63175]: _type = "Task" [ 1060.889940] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.897395] env[63175]: DEBUG oslo_vmware.api [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248614, 'name': ResetVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.076253] env[63175]: DEBUG nova.compute.manager [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1061.076546] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.077434] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f0d51c-7f0f-4586-a892-f9dfd7798edf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.085627] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.085874] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b231cf9-526a-4a18-a51e-18aad1204b5b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.092698] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1061.092698] env[63175]: value = "task-1248615" [ 1061.092698] env[63175]: _type = "Task" [ 1061.092698] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.100542] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.399714] env[63175]: DEBUG oslo_vmware.api [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248614, 'name': ResetVM_Task, 'duration_secs': 0.104842} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.400104] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Did hard reboot of VM {{(pid=63175) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1061.400346] env[63175]: DEBUG nova.compute.manager [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1061.401195] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a225d5b3-4775-4c7b-b1d6-5af1a1328899 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.604557] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248615, 'name': PowerOffVM_Task, 'duration_secs': 0.251925} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.604998] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.605110] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.605321] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dad45ad8-c708-4cd7-a5a4-a6ae835ef317 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.724960] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.725209] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.725457] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore1] 01025311-e8d6-4fc5-9e3f-e70e3aa3302d {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.725737] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15fbc728-6161-45e2-a6d4-fae37e3eb6f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.732636] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1061.732636] env[63175]: value = "task-1248617" [ 1061.732636] env[63175]: _type = "Task" [ 1061.732636] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.741470] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.913763] env[63175]: DEBUG oslo_concurrency.lockutils [None req-173934b4-afec-4792-8de6-6d22611e9d73 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.442s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.067654] env[63175]: DEBUG nova.compute.manager [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1062.067899] env[63175]: DEBUG nova.compute.manager [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing instance network info cache due to event network-changed-4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1062.069459] env[63175]: DEBUG oslo_concurrency.lockutils [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] Acquiring lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.069817] env[63175]: DEBUG oslo_concurrency.lockutils [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] Acquired lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.069817] env[63175]: DEBUG nova.network.neutron [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Refreshing network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.242936] env[63175]: DEBUG oslo_vmware.api [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191819} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.243363] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.243609] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.243841] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.244078] env[63175]: INFO nova.compute.manager [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1062.244387] env[63175]: DEBUG oslo.service.loopingcall [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.244641] env[63175]: DEBUG nova.compute.manager [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1062.244776] env[63175]: DEBUG nova.network.neutron [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1062.563474] env[63175]: DEBUG nova.compute.manager [req-3ec02846-1f13-480c-a20c-6c3e3a8c45c4 req-6ce9e9d4-dcab-4c47-9062-9b009631769e service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Received event network-vif-deleted-f0f01168-c8b5-446b-9224-a09b4e7843aa {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1062.563744] env[63175]: INFO nova.compute.manager [req-3ec02846-1f13-480c-a20c-6c3e3a8c45c4 req-6ce9e9d4-dcab-4c47-9062-9b009631769e service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Neutron deleted interface f0f01168-c8b5-446b-9224-a09b4e7843aa; detaching it from the instance and deleting it from the info cache [ 1062.563918] env[63175]: DEBUG nova.network.neutron [req-3ec02846-1f13-480c-a20c-6c3e3a8c45c4 req-6ce9e9d4-dcab-4c47-9062-9b009631769e service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.689115] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.689387] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.689598] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.689782] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.689952] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.692505] env[63175]: INFO nova.compute.manager [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Terminating instance [ 1062.781668] env[63175]: DEBUG nova.network.neutron [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updated VIF entry in instance network info cache for port 4bae9b2c-6be7-4054-99d0-6ff15c626fa0. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1062.782074] env[63175]: DEBUG nova.network.neutron [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [{"id": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "address": "fa:16:3e:e4:28:6a", "network": {"id": "b4ba845c-33ca-4188-a26b-cd95dbcdad01", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1928107081-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f4bb6b47581d42f586b3af144e291547", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "085fb0ff-9285-4f1d-a008-a14da4844357", "external-id": "nsx-vlan-transportzone-729", "segmentation_id": 729, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bae9b2c-6b", "ovs_interfaceid": "4bae9b2c-6be7-4054-99d0-6ff15c626fa0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.036674] env[63175]: DEBUG nova.network.neutron [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.066742] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9c63c12-6920-4d3f-b87d-9e097283a87b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.076987] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8411dcc-e9a3-43ba-9eb0-531ba2c16939 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.102055] env[63175]: DEBUG nova.compute.manager [req-3ec02846-1f13-480c-a20c-6c3e3a8c45c4 req-6ce9e9d4-dcab-4c47-9062-9b009631769e service nova] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Detach interface failed, port_id=f0f01168-c8b5-446b-9224-a09b4e7843aa, reason: Instance 01025311-e8d6-4fc5-9e3f-e70e3aa3302d could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1063.196343] env[63175]: DEBUG nova.compute.manager [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1063.196656] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1063.197698] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a643f825-2d37-4e38-a4ab-eb6579d8f9bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.206391] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1063.206671] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-608ecc67-b5dc-42f9-8771-45435d66ba36 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.215347] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1063.215347] env[63175]: value = "task-1248618" [ 1063.215347] env[63175]: _type = "Task" [ 1063.215347] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.223810] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248618, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.286757] env[63175]: DEBUG oslo_concurrency.lockutils [req-b593c943-370f-4587-9baa-873b8a3fc4ca req-24af7f75-ab5c-4f02-812e-df2e638dc68d service nova] Releasing lock "refresh_cache-f99ad0a9-5a93-4033-9c57-20bc7762a48c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.539596] env[63175]: INFO nova.compute.manager [-] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Took 1.29 seconds to deallocate network for instance. [ 1063.726063] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248618, 'name': PowerOffVM_Task, 'duration_secs': 0.184616} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.726405] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.726596] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1063.726852] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-632519b9-3214-43bc-af6e-4b0d52937ad2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.798727] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1063.800023] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1063.800023] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleting the datastore file [datastore1] f99ad0a9-5a93-4033-9c57-20bc7762a48c {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.800023] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a57ad038-7cc1-4994-a1d8-072dc613e375 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.806345] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1063.806345] env[63175]: value = "task-1248620" [ 1063.806345] env[63175]: _type = "Task" [ 1063.806345] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.813875] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248620, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.046027] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.046393] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.046653] env[63175]: DEBUG nova.objects.instance [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 01025311-e8d6-4fc5-9e3f-e70e3aa3302d {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.316440] env[63175]: DEBUG oslo_vmware.api [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248620, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165024} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.316688] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.316881] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1064.317077] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1064.317275] env[63175]: INFO nova.compute.manager [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1064.317628] env[63175]: DEBUG oslo.service.loopingcall [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.317741] env[63175]: DEBUG nova.compute.manager [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1064.317802] env[63175]: DEBUG nova.network.neutron [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1064.589916] env[63175]: DEBUG nova.compute.manager [req-38917d55-c7a0-4eef-a447-0005719f30e0 req-3701a357-0657-401a-bab2-5b59e9fc753c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Received event network-vif-deleted-4bae9b2c-6be7-4054-99d0-6ff15c626fa0 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1064.590142] env[63175]: INFO nova.compute.manager [req-38917d55-c7a0-4eef-a447-0005719f30e0 req-3701a357-0657-401a-bab2-5b59e9fc753c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Neutron deleted interface 4bae9b2c-6be7-4054-99d0-6ff15c626fa0; detaching it from the instance and deleting it from the info cache [ 1064.590401] env[63175]: DEBUG nova.network.neutron [req-38917d55-c7a0-4eef-a447-0005719f30e0 req-3701a357-0657-401a-bab2-5b59e9fc753c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.640826] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006ef74c-0339-4f84-ac95-3f9f6e85da8c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.650641] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.650868] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.652714] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16655bad-0e0e-468f-bd0e-ee547a13a23d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.683251] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a58215-33a1-4ad9-9cef-7ce1fad0e3e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.690505] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7653551c-e752-469e-8ff4-74446105eddf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.703599] env[63175]: DEBUG nova.compute.provider_tree [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.023779] env[63175]: DEBUG nova.network.neutron [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.095821] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8dbb72e-81bf-4204-9f7e-053827925561 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.106378] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98e1985-ce43-4104-b3d4-5aa7fc82301f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.132561] env[63175]: DEBUG nova.compute.manager [req-38917d55-c7a0-4eef-a447-0005719f30e0 req-3701a357-0657-401a-bab2-5b59e9fc753c service nova] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Detach interface failed, port_id=4bae9b2c-6be7-4054-99d0-6ff15c626fa0, reason: Instance f99ad0a9-5a93-4033-9c57-20bc7762a48c could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1065.157587] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1065.206825] env[63175]: DEBUG nova.scheduler.client.report [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1065.526072] env[63175]: INFO nova.compute.manager [-] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Took 1.21 seconds to deallocate network for instance. [ 1065.677789] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.711548] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.714019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.036s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.715481] env[63175]: INFO nova.compute.claims [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.729807] env[63175]: INFO nova.scheduler.client.report [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 01025311-e8d6-4fc5-9e3f-e70e3aa3302d [ 1066.032451] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.237420] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d6b2b4fa-2a12-491f-8313-50e4f4200e49 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "01025311-e8d6-4fc5-9e3f-e70e3aa3302d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.668s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.791613] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f27474-c992-489a-9ed1-a2aef816bbcd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.799225] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7229bc1a-20b8-4ec0-84e9-c65f3752f157 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.828983] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31522971-1af9-45c8-b620-14aabb6ed043 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.836646] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30387d1-06ea-4d7f-9729-815a7edbb0f5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.852529] env[63175]: DEBUG nova.compute.provider_tree [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.355779] env[63175]: DEBUG nova.scheduler.client.report [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1067.669451] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "45a4df18-72bf-40c5-a351-827221dd44db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.669692] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.860579] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.147s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.861134] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1067.863746] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.831s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.863977] env[63175]: DEBUG nova.objects.instance [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lazy-loading 'resources' on Instance uuid f99ad0a9-5a93-4033-9c57-20bc7762a48c {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.173929] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1068.367196] env[63175]: DEBUG nova.compute.utils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1068.371242] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Not allocating networking since 'none' was specified. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1068.440434] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46284066-f716-4389-8695-d24a2942c648 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.448200] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd7f8e9-6b80-4bda-be00-fc87ffa41649 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.478779] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5ae21c-1fd3-4fea-a15d-856b08fc5c88 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.486124] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fbea09-be47-4be3-9960-2f21736172a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.499083] env[63175]: DEBUG nova.compute.provider_tree [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.697908] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.872089] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1069.001816] env[63175]: DEBUG nova.scheduler.client.report [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1069.507069] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.643s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.509145] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.811s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.510600] env[63175]: INFO nova.compute.claims [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.526041] env[63175]: INFO nova.scheduler.client.report [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleted allocations for instance f99ad0a9-5a93-4033-9c57-20bc7762a48c [ 1069.882104] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1069.911381] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.911804] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.912117] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.912438] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.912707] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.912971] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.913352] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.913632] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.913928] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.914238] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.914551] env[63175]: DEBUG nova.virt.hardware [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.915911] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e52d13-5351-44fa-be28-cabfc8a65068 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.927811] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57980c2-abcf-485c-8170-87d798e768c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.950414] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.960534] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Creating folder: Project (5017a80ce48b433faa828cda1b567820). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1069.961019] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb3361cd-70fe-4425-981e-d52825830972 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.974853] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Created folder: Project (5017a80ce48b433faa828cda1b567820) in parent group-v268956. [ 1069.975043] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Creating folder: Instances. Parent ref: group-v269106. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1069.975421] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bad3bb7-ac51-4907-a760-2225dd463865 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.986398] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Created folder: Instances in parent group-v269106. [ 1069.986769] env[63175]: DEBUG oslo.service.loopingcall [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.987077] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1069.987383] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6716f37f-326d-4a47-84a8-52e078641af6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.011759] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.011759] env[63175]: value = "task-1248623" [ 1070.011759] env[63175]: _type = "Task" [ 1070.011759] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.025373] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248623, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.035132] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c135b1d0-8260-4225-83f1-93387275c0ec tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "f99ad0a9-5a93-4033-9c57-20bc7762a48c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.346s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.525587] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248623, 'name': CreateVM_Task, 'duration_secs': 0.260638} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.526128] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.526645] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.526681] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.527058] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.527320] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce8fe57c-ded7-480f-8a34-891e0285e242 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.531733] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1070.531733] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523c2297-e333-a8d4-9b7a-9802db3e6d55" [ 1070.531733] env[63175]: _type = "Task" [ 1070.531733] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.539487] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523c2297-e333-a8d4-9b7a-9802db3e6d55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.596499] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456a9c83-da72-41cc-b3a4-c7e479bd9257 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.605075] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df36326-a5a5-4bd2-9f3d-7e621ca56c0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.635472] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b34291e-3a05-4af3-8b48-c1585582273b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.643461] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05938de9-7c22-4c56-ac03-c674eccc8a8d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.657754] env[63175]: DEBUG nova.compute.provider_tree [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.044574] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523c2297-e333-a8d4-9b7a-9802db3e6d55, 'name': SearchDatastore_Task, 'duration_secs': 0.009625} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.044828] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.045076] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.045314] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.045504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.045687] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.045953] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5d6b0aa-fc81-4d29-a74d-569405a82bcc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.053833] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.054023] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.054692] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c59c9278-b234-46a4-885a-b30e6bc162c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.059413] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1071.059413] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bee467-b75a-ca6b-2fd3-d5dd52f0a570" [ 1071.059413] env[63175]: _type = "Task" [ 1071.059413] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.066747] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bee467-b75a-ca6b-2fd3-d5dd52f0a570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.164458] env[63175]: DEBUG nova.scheduler.client.report [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1071.169125] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "41b0611d-7249-4e81-bace-30418ca7478a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.169432] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.169734] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "41b0611d-7249-4e81-bace-30418ca7478a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.169998] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.170259] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.172673] env[63175]: INFO nova.compute.manager [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Terminating instance [ 1071.572860] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bee467-b75a-ca6b-2fd3-d5dd52f0a570, 'name': SearchDatastore_Task, 'duration_secs': 0.007868} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.573868] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4629746a-8262-4f5c-92bf-ead1fe0ffdf6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.580489] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1071.580489] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5204dc11-7a0e-e61c-c986-de4dc7a7f462" [ 1071.580489] env[63175]: _type = "Task" [ 1071.580489] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.589771] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5204dc11-7a0e-e61c-c986-de4dc7a7f462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.649565] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.649748] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Cleaning up deleted instances {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 1071.673413] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.674012] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1071.677053] env[63175]: DEBUG nova.compute.manager [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1071.677296] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1071.678399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef8b3d7-98cd-4e46-9607-ad61958f8b81 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.686483] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.686730] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b60475aa-d881-4256-b229-5623fc70ac0f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.693377] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1071.693377] env[63175]: value = "task-1248624" [ 1071.693377] env[63175]: _type = "Task" [ 1071.693377] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.701678] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.090638] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5204dc11-7a0e-e61c-c986-de4dc7a7f462, 'name': SearchDatastore_Task, 'duration_secs': 0.009945} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.090898] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.091175] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.091426] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ad4cb0e-8a28-4592-bf22-48a810a5e0c9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.097682] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1072.097682] env[63175]: value = "task-1248625" [ 1072.097682] env[63175]: _type = "Task" [ 1072.097682] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.104651] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.164262] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] There are 49 instances to clean {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 1072.164574] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 01025311-e8d6-4fc5-9e3f-e70e3aa3302d] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1072.178967] env[63175]: DEBUG nova.compute.utils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1072.180326] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1072.180518] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1072.202987] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248624, 'name': PowerOffVM_Task, 'duration_secs': 0.180061} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.203751] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.203961] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1072.204246] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c00f0ee2-6a87-43d1-8ea8-d8d0cd14509e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.218519] env[63175]: DEBUG nova.policy [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1072.286134] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1072.286519] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1072.286762] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleting the datastore file [datastore2] 41b0611d-7249-4e81-bace-30418ca7478a {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.289206] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46c4a3c7-5747-4529-beb1-029b7d616dd0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.295996] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for the task: (returnval){ [ 1072.295996] env[63175]: value = "task-1248627" [ 1072.295996] env[63175]: _type = "Task" [ 1072.295996] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.306480] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248627, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.535115] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Successfully created port: 8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.609869] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.414626} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.610223] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.610374] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.610624] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52d133bc-2273-4743-b380-ab3fd8c588ba {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.617139] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1072.617139] env[63175]: value = "task-1248628" [ 1072.617139] env[63175]: _type = "Task" [ 1072.617139] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.624343] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248628, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.668192] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: f99ad0a9-5a93-4033-9c57-20bc7762a48c] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1072.683999] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1072.806099] env[63175]: DEBUG oslo_vmware.api [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Task: {'id': task-1248627, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263696} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.806430] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.806663] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1072.806927] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1072.807106] env[63175]: INFO nova.compute.manager [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1072.807403] env[63175]: DEBUG oslo.service.loopingcall [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.807660] env[63175]: DEBUG nova.compute.manager [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1072.807867] env[63175]: DEBUG nova.network.neutron [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1073.056117] env[63175]: DEBUG nova.compute.manager [req-60c64c46-9083-4b46-983c-5d42837a2bb4 req-c61d9c97-6272-4356-90ab-e6511e8d02c9 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Received event network-vif-deleted-7e1877cb-140e-4619-ae3b-5de462079aeb {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1073.056415] env[63175]: INFO nova.compute.manager [req-60c64c46-9083-4b46-983c-5d42837a2bb4 req-c61d9c97-6272-4356-90ab-e6511e8d02c9 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Neutron deleted interface 7e1877cb-140e-4619-ae3b-5de462079aeb; detaching it from the instance and deleting it from the info cache [ 1073.056506] env[63175]: DEBUG nova.network.neutron [req-60c64c46-9083-4b46-983c-5d42837a2bb4 req-c61d9c97-6272-4356-90ab-e6511e8d02c9 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.126650] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248628, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108316} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.126923] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.127700] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1105aae7-e800-432d-aa1d-2ea36884974c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.154287] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.154562] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67a1f767-bf3d-4d08-9d91-4a8dee176695 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.170820] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 6d2786d6-55e4-4d82-8a8c-b725449166a1] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1073.174357] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1073.174357] env[63175]: value = "task-1248629" [ 1073.174357] env[63175]: _type = "Task" [ 1073.174357] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.182148] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.534932] env[63175]: DEBUG nova.network.neutron [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.559493] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22f42ad2-319e-4141-b522-8d6f3885003e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.568919] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9be033-dc74-442b-9d36-c90912ceefff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.594752] env[63175]: DEBUG nova.compute.manager [req-60c64c46-9083-4b46-983c-5d42837a2bb4 req-c61d9c97-6272-4356-90ab-e6511e8d02c9 service nova] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Detach interface failed, port_id=7e1877cb-140e-4619-ae3b-5de462079aeb, reason: Instance 41b0611d-7249-4e81-bace-30418ca7478a could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1073.674134] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: c08b4b19-6c3d-47fa-be5b-d4cf1675e4a5] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1073.685377] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248629, 'name': ReconfigVM_Task, 'duration_secs': 0.388409} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.685700] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.686315] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2984862-6b2b-4ab1-a0cf-93907547b38f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.692062] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1073.695346] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1073.695346] env[63175]: value = "task-1248630" [ 1073.695346] env[63175]: _type = "Task" [ 1073.695346] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.704953] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248630, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.720152] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.720437] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.720621] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.720877] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.721294] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.721469] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.721699] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.721879] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.722074] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.722250] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.722431] env[63175]: DEBUG nova.virt.hardware [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.723620] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9021595-f659-4df5-bf22-5e13c3f2146d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.731821] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce573737-7aed-47cd-8934-234ad26372c2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.921639] env[63175]: DEBUG nova.compute.manager [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Received event network-vif-plugged-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1073.921877] env[63175]: DEBUG oslo_concurrency.lockutils [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] Acquiring lock "45a4df18-72bf-40c5-a351-827221dd44db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.922110] env[63175]: DEBUG oslo_concurrency.lockutils [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] Lock "45a4df18-72bf-40c5-a351-827221dd44db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.922298] env[63175]: DEBUG oslo_concurrency.lockutils [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] Lock "45a4df18-72bf-40c5-a351-827221dd44db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.922500] env[63175]: DEBUG nova.compute.manager [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] No waiting events found dispatching network-vif-plugged-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1073.922731] env[63175]: WARNING nova.compute.manager [req-44256368-f3e1-4e95-ac1e-815f74a5f0fd req-aed5eb27-3d78-4a30-a2af-672b71aae878 service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Received unexpected event network-vif-plugged-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 for instance with vm_state building and task_state spawning. [ 1074.003383] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Successfully updated port: 8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.037923] env[63175]: INFO nova.compute.manager [-] [instance: 41b0611d-7249-4e81-bace-30418ca7478a] Took 1.23 seconds to deallocate network for instance. [ 1074.180375] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 19ec0aa6-3078-419d-849d-1bb371325a87] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1074.204763] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248630, 'name': Rename_Task, 'duration_secs': 0.219153} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.205032] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.205273] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26095d0a-1d62-42ab-94f5-97c16df516c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.212143] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1074.212143] env[63175]: value = "task-1248631" [ 1074.212143] env[63175]: _type = "Task" [ 1074.212143] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.219537] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.506251] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.506468] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.506645] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1074.546024] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.546286] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.546522] env[63175]: DEBUG nova.objects.instance [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lazy-loading 'resources' on Instance uuid 41b0611d-7249-4e81-bace-30418ca7478a {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.683431] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: b8c2febb-7024-49de-b34c-1c8ee492b39f] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1074.724118] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248631, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.039481] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1075.127447] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276140fb-bf9b-40e5-a950-b5cb939d84c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.135804] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56736943-90c1-4226-938e-ae5a1461febb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.169022] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8562581a-21d7-4a6a-87fb-aeaa96ec25df {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.177183] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964440d3-2000-4048-8de0-f5beaa776ce0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.191074] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 8a4877b2-6284-4727-b1fc-cf30bbc65e36] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1075.193043] env[63175]: DEBUG nova.compute.provider_tree [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.210190] env[63175]: DEBUG nova.network.neutron [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Updating instance_info_cache with network_info: [{"id": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "address": "fa:16:3e:58:f5:e0", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ea56d43-c9", "ovs_interfaceid": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.224019] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248631, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.695663] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 9a7328a7-3cfa-4bf0-b9e9-e6231de1d4e0] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1075.697490] env[63175]: DEBUG nova.scheduler.client.report [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1075.713413] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.713512] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Instance network_info: |[{"id": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "address": "fa:16:3e:58:f5:e0", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ea56d43-c9", "ovs_interfaceid": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1075.714403] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:f5:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ea56d43-c9b6-45a9-aaf8-98da64c6ee80', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.724649] env[63175]: DEBUG oslo.service.loopingcall [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.725232] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.728569] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-863a0acf-73ca-45c3-8d9b-3b41db0fc795 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.754836] env[63175]: DEBUG oslo_vmware.api [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248631, 'name': PowerOnVM_Task, 'duration_secs': 1.086138} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.756132] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.756408] env[63175]: INFO nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Took 5.87 seconds to spawn the instance on the hypervisor. [ 1075.756605] env[63175]: DEBUG nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1075.756904] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.756904] env[63175]: value = "task-1248632" [ 1075.756904] env[63175]: _type = "Task" [ 1075.756904] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.757724] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c92759-a702-476a-bfc1-aef401c58d65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.772479] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.951431] env[63175]: DEBUG nova.compute.manager [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Received event network-changed-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1075.951710] env[63175]: DEBUG nova.compute.manager [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Refreshing instance network info cache due to event network-changed-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1075.952043] env[63175]: DEBUG oslo_concurrency.lockutils [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] Acquiring lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.952180] env[63175]: DEBUG oslo_concurrency.lockutils [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] Acquired lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.952412] env[63175]: DEBUG nova.network.neutron [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Refreshing network info cache for port 8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1076.201964] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 48448a68-6b7e-488b-a7f2-9f910a2f6765] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1076.204464] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.223228] env[63175]: INFO nova.scheduler.client.report [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Deleted allocations for instance 41b0611d-7249-4e81-bace-30418ca7478a [ 1076.270520] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.278377] env[63175]: INFO nova.compute.manager [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Took 10.62 seconds to build instance. [ 1076.657829] env[63175]: DEBUG nova.network.neutron [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Updated VIF entry in instance network info cache for port 8ea56d43-c9b6-45a9-aaf8-98da64c6ee80. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1076.658232] env[63175]: DEBUG nova.network.neutron [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Updating instance_info_cache with network_info: [{"id": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "address": "fa:16:3e:58:f5:e0", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ea56d43-c9", "ovs_interfaceid": "8ea56d43-c9b6-45a9-aaf8-98da64c6ee80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.707326] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 63be6b3c-2535-4c26-9a15-9c2445b049e8] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1076.730791] env[63175]: DEBUG oslo_concurrency.lockutils [None req-43fd837c-e7b5-488d-8498-6765ae09c086 tempest-SecurityGroupsTestJSON-323758155 tempest-SecurityGroupsTestJSON-323758155-project-member] Lock "41b0611d-7249-4e81-bace-30418ca7478a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.561s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.771093] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.779901] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1df15a53-9e95-47ce-b616-20b30c8fb8ba tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.129s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.948487] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.948813] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.160673] env[63175]: DEBUG oslo_concurrency.lockutils [req-a292bc33-a8c2-4ba0-bf23-7ac0b36b1065 req-c59c7051-adeb-4bc6-afef-3fe75d916d2c service nova] Releasing lock "refresh_cache-45a4df18-72bf-40c5-a351-827221dd44db" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.210463] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 572a3821-7436-487a-a053-3819411de57e] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1077.271938] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 25%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.452312] env[63175]: DEBUG nova.compute.utils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.494024] env[63175]: INFO nova.compute.manager [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Rebuilding instance [ 1077.534848] env[63175]: DEBUG nova.compute.manager [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1077.535720] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f8e628-4aba-4655-9477-d11976478f6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.713487] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 089ce349-ac2d-4752-a41a-562b4acd2e73] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1077.772519] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.955618] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.217439] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 2eec3ef1-ea7c-4e2b-b31c-77529c2a76f5] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1078.272741] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.550202] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.550538] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c85e33cf-19a4-4e2b-82ee-86a8161522ef {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.558106] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1078.558106] env[63175]: value = "task-1248633" [ 1078.558106] env[63175]: _type = "Task" [ 1078.558106] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.566352] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.720897] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0633c8e9-1c18-4047-ae5d-012d715e03df] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1078.773545] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248632, 'name': CreateVM_Task, 'duration_secs': 2.882291} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.773751] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.774502] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.774721] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.775055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.775358] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef36f770-ab00-4de4-b769-530f623c0606 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.779533] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1078.779533] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d63a1-43e2-8c10-a781-cb6dadc57103" [ 1078.779533] env[63175]: _type = "Task" [ 1078.779533] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.787111] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d63a1-43e2-8c10-a781-cb6dadc57103, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.027071] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.027410] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.027669] env[63175]: INFO nova.compute.manager [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Attaching volume 959e1d2c-e8ac-4673-94e7-e57f912db488 to /dev/sdb [ 1079.064354] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd2ab8e-e994-47c4-bcf2-d9010ef5074c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.073766] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248633, 'name': PowerOffVM_Task, 'duration_secs': 0.132499} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.075403] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.075679] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1079.076478] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13505c7-fed5-434c-8d02-784c8a42c2b5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.079480] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4dd648-af2a-4dac-8888-9fbde0a8ace8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.086498] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1079.086753] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0333e03-a885-4463-bb82-662f8a00d0cf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.094482] env[63175]: DEBUG nova.virt.block_device [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating existing volume attachment record: f34ddecb-bdca-424a-9153-462dd822ba05 {{(pid=63175) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1079.121129] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1079.121365] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1079.121588] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Deleting the datastore file [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.122735] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-291b7f2e-5810-48ba-b705-8eaa687d866a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.128149] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1079.128149] env[63175]: value = "task-1248635" [ 1079.128149] env[63175]: _type = "Task" [ 1079.128149] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.137560] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.224054] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 9057bbcd-9c7c-4fba-9c3c-3a7dbbb045db] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1079.289998] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]529d63a1-43e2-8c10-a781-cb6dadc57103, 'name': SearchDatastore_Task, 'duration_secs': 0.016159} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.290348] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.290586] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.290833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.290980] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.291190] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.291455] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf7c2525-253d-4920-95bf-cad855f96fb3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.299837] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.300039] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.300769] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea060aaf-a477-4538-85fd-74ee80066ad0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.306221] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1079.306221] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525b33ae-4b6d-99e8-3c78-e81f5dfc4beb" [ 1079.306221] env[63175]: _type = "Task" [ 1079.306221] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.314561] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525b33ae-4b6d-99e8-3c78-e81f5dfc4beb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.638578] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093677} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.638912] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.639184] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.639411] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.727944] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: ed618dba-1b54-4119-a317-4f64a64a59c9] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1079.816782] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525b33ae-4b6d-99e8-3c78-e81f5dfc4beb, 'name': SearchDatastore_Task, 'duration_secs': 0.009148} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.817666] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d65f8a-eb2b-4ab6-a1aa-f95bf959869c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.822890] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1079.822890] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52302a01-babf-50c6-37f8-a638ea203b74" [ 1079.822890] env[63175]: _type = "Task" [ 1079.822890] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.831113] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52302a01-babf-50c6-37f8-a638ea203b74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.233261] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: aad1653a-43a7-4637-acd5-7dbf76ea1ccc] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1080.333724] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52302a01-babf-50c6-37f8-a638ea203b74, 'name': SearchDatastore_Task, 'duration_secs': 0.009103} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.334482] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.334482] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 45a4df18-72bf-40c5-a351-827221dd44db/45a4df18-72bf-40c5-a351-827221dd44db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1080.334730] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b817a0c9-cae6-4c1d-98ce-b9e8c7ba70fa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.341136] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1080.341136] env[63175]: value = "task-1248637" [ 1080.341136] env[63175]: _type = "Task" [ 1080.341136] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.348698] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.685431] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.685800] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.685982] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.686366] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.686665] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.686883] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.687524] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.688239] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.688504] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.688757] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.688985] env[63175]: DEBUG nova.virt.hardware [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.690188] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0d182a-0d0f-4e32-8a3e-9e82615d9771 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.702143] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689bd1d8-fad6-4bc2-879e-1ead4e55e45c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.720727] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance VIF info [] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.728539] env[63175]: DEBUG oslo.service.loopingcall [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.728877] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1080.729273] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3827af1-ea0c-4d69-9fa3-c35535ada211 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.744825] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 43f8c7c8-d8cb-4b1c-b0b7-f16d0981e4b8] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1080.755031] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.755031] env[63175]: value = "task-1248638" [ 1080.755031] env[63175]: _type = "Task" [ 1080.755031] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.763285] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248638, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.852092] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248637, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.250062] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 8d3bd040-fa87-44b8-a6c1-fe6be21b0415] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1081.265814] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248638, 'name': CreateVM_Task, 'duration_secs': 0.284597} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.266042] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1081.267226] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.267226] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.267226] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.267393] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34fa1d0a-1e55-4017-8816-29d176039ae7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.274981] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1081.274981] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dd104e-07c0-7105-3d7d-25153977609c" [ 1081.274981] env[63175]: _type = "Task" [ 1081.274981] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.283567] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dd104e-07c0-7105-3d7d-25153977609c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.352909] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512753} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.353200] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 45a4df18-72bf-40c5-a351-827221dd44db/45a4df18-72bf-40c5-a351-827221dd44db.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1081.353421] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.353790] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55497c63-7fe5-4d5c-a61a-fab27e1f183d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.360618] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1081.360618] env[63175]: value = "task-1248640" [ 1081.360618] env[63175]: _type = "Task" [ 1081.360618] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.368851] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248640, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.753337] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0645fffe-9910-43af-af40-126592aefaff] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1081.785835] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52dd104e-07c0-7105-3d7d-25153977609c, 'name': SearchDatastore_Task, 'duration_secs': 0.022321} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.786150] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.786381] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.786655] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.786807] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.786989] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.787264] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b32f5af-b0d1-49a8-8584-ea0758c1bc00 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.796360] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.796544] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.797259] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-328a954b-e889-4579-b2a4-04558a7857f3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.802796] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1081.802796] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52472ca0-cd50-40ec-a185-cf5a075bbb6c" [ 1081.802796] env[63175]: _type = "Task" [ 1081.802796] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.810723] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52472ca0-cd50-40ec-a185-cf5a075bbb6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.871079] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248640, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087228} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.871079] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.871823] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb784f4-5f30-48c1-82b0-aa16ec790d18 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.896866] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 45a4df18-72bf-40c5-a351-827221dd44db/45a4df18-72bf-40c5-a351-827221dd44db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.897186] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb5faba3-6d07-413d-b55e-8abe985c2d09 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.919709] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1081.919709] env[63175]: value = "task-1248641" [ 1081.919709] env[63175]: _type = "Task" [ 1081.919709] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.925298] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248641, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.259553] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a4304579-f829-433e-a878-1050fac08c2e] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1082.318553] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52472ca0-cd50-40ec-a185-cf5a075bbb6c, 'name': SearchDatastore_Task, 'duration_secs': 0.011012} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.318553] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d64d058f-8c98-4f64-8f5c-84425c3e4440 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.321223] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1082.321223] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a16a38-8235-aaa7-ef4f-75324d53c15a" [ 1082.321223] env[63175]: _type = "Task" [ 1082.321223] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.329375] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a16a38-8235-aaa7-ef4f-75324d53c15a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.430097] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248641, 'name': ReconfigVM_Task, 'duration_secs': 0.350544} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.430389] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 45a4df18-72bf-40c5-a351-827221dd44db/45a4df18-72bf-40c5-a351-827221dd44db.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.431045] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c3c1400-e057-4238-bba6-14e47837aeeb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.437943] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1082.437943] env[63175]: value = "task-1248642" [ 1082.437943] env[63175]: _type = "Task" [ 1082.437943] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.446547] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248642, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.763616] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 9b72d607-be0f-4caa-b9e2-b25bbbe30f54] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1082.832202] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a16a38-8235-aaa7-ef4f-75324d53c15a, 'name': SearchDatastore_Task, 'duration_secs': 0.008616} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.832504] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.832772] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.833041] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85ac9d91-9cc1-4471-b6e4-dff817eddac4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.839784] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1082.839784] env[63175]: value = "task-1248643" [ 1082.839784] env[63175]: _type = "Task" [ 1082.839784] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.848421] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248643, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.949539] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248642, 'name': Rename_Task, 'duration_secs': 0.130865} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.949917] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.950376] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf38f947-88b9-492f-9988-15bfb550d691 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.957728] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1082.957728] env[63175]: value = "task-1248644" [ 1082.957728] env[63175]: _type = "Task" [ 1082.957728] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.967490] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.267196] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 5d25451b-69aa-4af7-be71-451fa0ac1e39] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1083.352481] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248643, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.472618] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248644, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.648936] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Volume attach. Driver type: vmdk {{(pid=63175) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1083.649178] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269110', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'name': 'volume-959e1d2c-e8ac-4673-94e7-e57f912db488', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1579e4ad-f754-46ea-a9af-c85947489dcd', 'attached_at': '', 'detached_at': '', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'serial': '959e1d2c-e8ac-4673-94e7-e57f912db488'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1083.650066] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e056493-0f3d-407d-802b-192c8b2df1f9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.666578] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a672bb96-a3d4-47c0-9cce-0d2efc9114ec {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.692361] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-959e1d2c-e8ac-4673-94e7-e57f912db488/volume-959e1d2c-e8ac-4673-94e7-e57f912db488.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.692680] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-397aa3f4-6127-4593-ba33-f4728579288e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.713023] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1083.713023] env[63175]: value = "task-1248645" [ 1083.713023] env[63175]: _type = "Task" [ 1083.713023] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.721545] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.773011] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 1b8b3fdf-642b-40cc-a8ba-07ecd5519a68] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1083.850771] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248643, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546155} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.851022] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.851255] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.851514] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0715ecb-ddc0-467e-96a0-679429de204a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.858553] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1083.858553] env[63175]: value = "task-1248646" [ 1083.858553] env[63175]: _type = "Task" [ 1083.858553] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.865329] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.969928] env[63175]: DEBUG oslo_vmware.api [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248644, 'name': PowerOnVM_Task, 'duration_secs': 0.722846} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.970235] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.970430] env[63175]: INFO nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Took 10.28 seconds to spawn the instance on the hypervisor. [ 1083.970619] env[63175]: DEBUG nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1083.971466] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fd152a-38fa-491f-8940-dbbc960a2b28 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.089037] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "225096ef-500f-490f-b3c0-def1c46ed17c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.089159] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.222794] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248645, 'name': ReconfigVM_Task, 'duration_secs': 0.474381} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.223100] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-959e1d2c-e8ac-4673-94e7-e57f912db488/volume-959e1d2c-e8ac-4673-94e7-e57f912db488.vmdk or device None with type thin {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.228268] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fa03843-a056-48b1-bd9c-0b57703dea42 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.243507] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1084.243507] env[63175]: value = "task-1248647" [ 1084.243507] env[63175]: _type = "Task" [ 1084.243507] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.253135] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248647, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.275712] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: f82caee9-bb2f-4c07-b7f0-018ac82c4066] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1084.366721] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109266} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.367017] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.367854] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a852c917-8867-49f0-a1b8-ca453765bffa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.386870] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.387125] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ee1014d-9864-4ab3-8618-478042e5d2c7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.405446] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1084.405446] env[63175]: value = "task-1248648" [ 1084.405446] env[63175]: _type = "Task" [ 1084.405446] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.412839] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.492808] env[63175]: INFO nova.compute.manager [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Took 15.81 seconds to build instance. [ 1084.591591] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1084.760324] env[63175]: DEBUG oslo_vmware.api [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248647, 'name': ReconfigVM_Task, 'duration_secs': 0.135549} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.760851] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269110', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'name': 'volume-959e1d2c-e8ac-4673-94e7-e57f912db488', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1579e4ad-f754-46ea-a9af-c85947489dcd', 'attached_at': '', 'detached_at': '', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'serial': '959e1d2c-e8ac-4673-94e7-e57f912db488'} {{(pid=63175) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1084.780946] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: f37d34d4-b13e-4ce0-b010-3a06feab6324] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1084.915782] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248648, 'name': ReconfigVM_Task, 'duration_secs': 0.265029} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.916622] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0/86fd8d46-fcec-4d07-be79-d307d1fedcf0.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.916839] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87b2f4f9-c251-4331-b8e3-c405f93b6346 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.923210] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1084.923210] env[63175]: value = "task-1248649" [ 1084.923210] env[63175]: _type = "Task" [ 1084.923210] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.931462] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248649, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.997696] env[63175]: DEBUG oslo_concurrency.lockutils [None req-08c21b08-b08d-4a08-9ab0-9c50f5707b67 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.328s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.182111] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.182111] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.182111] env[63175]: INFO nova.compute.claims [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.285546] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 26ab94cc-c5ff-4d69-9d27-f01ca8038b30] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1085.433819] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248649, 'name': Rename_Task, 'duration_secs': 0.13462} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.434123] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.434372] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c1a09e6-02d7-4b45-829c-d6540e6f3fee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.440638] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1085.440638] env[63175]: value = "task-1248650" [ 1085.440638] env[63175]: _type = "Task" [ 1085.440638] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.449810] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.787557] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0bd5141f-0b92-4d27-8d51-023ab1096e78] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1085.816800] env[63175]: DEBUG nova.objects.instance [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid 1579e4ad-f754-46ea-a9af-c85947489dcd {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.950942] env[63175]: DEBUG oslo_vmware.api [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248650, 'name': PowerOnVM_Task, 'duration_secs': 0.482191} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.951964] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.951964] env[63175]: DEBUG nova.compute.manager [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1085.952328] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d292834f-d12e-4cfc-ab2b-c6d92e491d77 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.084905] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.085182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.281220] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a73a54-ff4d-4f3d-ae19-8ef2a99923d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.289636] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cff409f-ca8c-4503-a520-94e141b4d42b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.293779] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 4bd91412-c41b-41a1-a648-6b905d826ee3] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1086.325705] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d2ad2bd7-c15b-4c3b-99b4-93eb1aea8625 tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.298s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.327065] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d78bc3c-0079-486c-82a2-4ffa7338d6bd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.335124] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2480a872-f760-42e1-88f1-8c380954b6cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.349800] env[63175]: DEBUG nova.compute.provider_tree [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.468364] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.540486] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.540753] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.587163] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1086.796798] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 54440032-4d9b-41d4-9ef2-5a79a4224fa6] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1086.853161] env[63175]: DEBUG nova.scheduler.client.report [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1087.043943] env[63175]: INFO nova.compute.manager [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Detaching volume 959e1d2c-e8ac-4673-94e7-e57f912db488 [ 1087.078658] env[63175]: INFO nova.virt.block_device [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Attempting to driver detach volume 959e1d2c-e8ac-4673-94e7-e57f912db488 from mountpoint /dev/sdb [ 1087.078958] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Volume detach. Driver type: vmdk {{(pid=63175) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.079184] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269110', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'name': 'volume-959e1d2c-e8ac-4673-94e7-e57f912db488', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1579e4ad-f754-46ea-a9af-c85947489dcd', 'attached_at': '', 'detached_at': '', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'serial': '959e1d2c-e8ac-4673-94e7-e57f912db488'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.080076] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4cbde9-f33a-49c4-ae1c-b0d1441f26b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.103782] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0568d3c0-4d70-4fca-97e8-37581a3adb6a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.112574] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da611c0-76fe-4373-bb49-42f199365db6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.134709] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.135469] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2375681c-00ee-469c-95a0-a9f774404641 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.149986] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] The volume has not been displaced from its original location: [datastore1] volume-959e1d2c-e8ac-4673-94e7-e57f912db488/volume-959e1d2c-e8ac-4673-94e7-e57f912db488.vmdk. No consolidation needed. {{(pid=63175) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1087.155325] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.155644] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a31c61b4-480f-4a22-9305-5a479453a0d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.173645] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1087.173645] env[63175]: value = "task-1248651" [ 1087.173645] env[63175]: _type = "Task" [ 1087.173645] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.184865] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.256458] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.256920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.257273] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.257798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.257798] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.259743] env[63175]: INFO nova.compute.manager [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Terminating instance [ 1087.300359] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a460926e-9637-40aa-bb30-e3890a441e03] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1087.358144] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.358640] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1087.361390] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.893s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.361576] env[63175]: DEBUG nova.objects.instance [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63175) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1087.683491] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248651, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.762911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "refresh_cache-86fd8d46-fcec-4d07-be79-d307d1fedcf0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.763261] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquired lock "refresh_cache-86fd8d46-fcec-4d07-be79-d307d1fedcf0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.763466] env[63175]: DEBUG nova.network.neutron [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.803667] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 46baf816-eff2-400c-b81c-0d5f3ce8d01e] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1087.866030] env[63175]: DEBUG nova.compute.utils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1087.867340] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1087.867516] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1087.914652] env[63175]: DEBUG nova.policy [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ab31d05840845c3a17005e4059ea557', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8efcd3690254590a3f636bba93809a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1088.150320] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Successfully created port: 01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1088.183559] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248651, 'name': ReconfigVM_Task, 'duration_secs': 0.628465} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.183892] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=63175) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.192164] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-518039d4-e21e-4ff5-8d95-540f482935d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.213486] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1088.213486] env[63175]: value = "task-1248652" [ 1088.213486] env[63175]: _type = "Task" [ 1088.213486] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.221009] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248652, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.282201] env[63175]: DEBUG nova.network.neutron [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.307434] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 5a2d540c-cea7-4db2-a11a-6b7dd0f7e82f] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1088.333360] env[63175]: DEBUG nova.network.neutron [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.373520] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1088.380368] env[63175]: DEBUG oslo_concurrency.lockutils [None req-4d015916-4e7d-4aeb-a590-1232bb2b62d1 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.381668] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.247s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.383311] env[63175]: INFO nova.compute.claims [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1088.723680] env[63175]: DEBUG oslo_vmware.api [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248652, 'name': ReconfigVM_Task, 'duration_secs': 0.144642} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.723986] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269110', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'name': 'volume-959e1d2c-e8ac-4673-94e7-e57f912db488', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1579e4ad-f754-46ea-a9af-c85947489dcd', 'attached_at': '', 'detached_at': '', 'volume_id': '959e1d2c-e8ac-4673-94e7-e57f912db488', 'serial': '959e1d2c-e8ac-4673-94e7-e57f912db488'} {{(pid=63175) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1088.809310] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: cf12aac3-edbb-48eb-b431-70187a4ecda3] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1088.835502] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Releasing lock "refresh_cache-86fd8d46-fcec-4d07-be79-d307d1fedcf0" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.835969] env[63175]: DEBUG nova.compute.manager [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1088.836193] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1088.837273] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a199ef3-b6cc-4823-a561-d77e1e424e9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.845135] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1088.845355] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27517fea-0653-4d60-aa08-b4ebe8bc4ed4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.851017] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1088.851017] env[63175]: value = "task-1248653" [ 1088.851017] env[63175]: _type = "Task" [ 1088.851017] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.858172] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.268441] env[63175]: DEBUG nova.objects.instance [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'flavor' on Instance uuid 1579e4ad-f754-46ea-a9af-c85947489dcd {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.312115] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 780989ee-98a1-4e99-9014-45de6921d4bc] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1089.361433] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248653, 'name': PowerOffVM_Task, 'duration_secs': 0.119124} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.361657] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.361830] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.362096] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4158b5d5-d469-4a50-8a64-fb6692736c4b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.384480] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.384708] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.384899] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Deleting the datastore file [datastore1] 86fd8d46-fcec-4d07-be79-d307d1fedcf0 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.385180] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81a561ec-2577-4924-b90f-77ee2afd563c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.387906] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1089.396938] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for the task: (returnval){ [ 1089.396938] env[63175]: value = "task-1248655" [ 1089.396938] env[63175]: _type = "Task" [ 1089.396938] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.405399] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.413478] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.413694] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.413858] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.414074] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.414216] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.414366] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.414579] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.414747] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.414919] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.415108] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.415293] env[63175]: DEBUG nova.virt.hardware [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.416173] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca306ea-0cf4-4122-b48e-c84c90f28582 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.426262] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7451e3cb-b0da-490f-8c7e-8d1bcfb9a2be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.482648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e28045-4ed9-49ed-b699-517032f82b9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.489979] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbe05c2-58ee-4689-9eb8-2e3fa8b77d55 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.521288] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c9a450-99a9-4f8d-872e-da3ff004f73d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.529171] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783ae31e-d9ab-42ce-9e08-f79f7d4d4ba7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.543367] env[63175]: DEBUG nova.compute.provider_tree [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.546332] env[63175]: DEBUG nova.compute.manager [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Received event network-vif-plugged-01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1089.546544] env[63175]: DEBUG oslo_concurrency.lockutils [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] Acquiring lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.546792] env[63175]: DEBUG oslo_concurrency.lockutils [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] Lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.547037] env[63175]: DEBUG oslo_concurrency.lockutils [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] Lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.547155] env[63175]: DEBUG nova.compute.manager [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] No waiting events found dispatching network-vif-plugged-01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1089.547420] env[63175]: WARNING nova.compute.manager [req-70b486fc-a226-48a7-96a0-c9af835370b2 req-57a888ab-46bd-4854-8655-e0ce547cf335 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Received unexpected event network-vif-plugged-01dd3e04-4265-4a89-b499-1e99878db549 for instance with vm_state building and task_state spawning. [ 1089.630963] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Successfully updated port: 01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1089.815766] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: e1e4f169-000c-4e9c-8ef5-aa4b4989eb44] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1089.907462] env[63175]: DEBUG oslo_vmware.api [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Task: {'id': task-1248655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41896} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.907727] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1089.907921] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1089.908114] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.908296] env[63175]: INFO nova.compute.manager [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1089.908538] env[63175]: DEBUG oslo.service.loopingcall [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.908864] env[63175]: DEBUG nova.compute.manager [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1089.908864] env[63175]: DEBUG nova.network.neutron [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.928749] env[63175]: DEBUG nova.network.neutron [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1090.049410] env[63175]: DEBUG nova.scheduler.client.report [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1090.132767] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.132920] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquired lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.133089] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1090.276529] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d063bd37-5581-4243-968d-a9c744b6e02b tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.736s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.319175] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 4d07a537-7df6-4659-8760-bf7e7925da25] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1090.430905] env[63175]: DEBUG nova.network.neutron [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.554432] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.554749] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1090.665336] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1090.788737] env[63175]: DEBUG nova.network.neutron [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Updating instance_info_cache with network_info: [{"id": "01dd3e04-4265-4a89-b499-1e99878db549", "address": "fa:16:3e:26:92:8c", "network": {"id": "36e1781f-25ae-41d0-bb2d-66ce72d0d672", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-937219789-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8efcd3690254590a3f636bba93809a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01dd3e04-42", "ovs_interfaceid": "01dd3e04-4265-4a89-b499-1e99878db549", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.794651] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.794896] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.795119] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.795368] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.795688] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.797492] env[63175]: INFO nova.compute.manager [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Terminating instance [ 1090.821904] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 72d131a0-2617-49a1-8aff-897908929bb0] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1090.934754] env[63175]: INFO nova.compute.manager [-] [instance: 86fd8d46-fcec-4d07-be79-d307d1fedcf0] Took 1.03 seconds to deallocate network for instance. [ 1091.059472] env[63175]: DEBUG nova.compute.utils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1091.060934] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1091.061116] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1091.100818] env[63175]: DEBUG nova.policy [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1091.291839] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Releasing lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.292215] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Instance network_info: |[{"id": "01dd3e04-4265-4a89-b499-1e99878db549", "address": "fa:16:3e:26:92:8c", "network": {"id": "36e1781f-25ae-41d0-bb2d-66ce72d0d672", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-937219789-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8efcd3690254590a3f636bba93809a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01dd3e04-42", "ovs_interfaceid": "01dd3e04-4265-4a89-b499-1e99878db549", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1091.292659] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:92:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04ccbc7a-cf8d-4ea2-8411-291a1e27df7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01dd3e04-4265-4a89-b499-1e99878db549', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1091.300197] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Creating folder: Project (b8efcd3690254590a3f636bba93809a6). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1091.300482] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d5bb8f7-daa7-49cd-be97-95cf07479f4c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.302696] env[63175]: DEBUG nova.compute.manager [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1091.302900] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.303689] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873953c2-b525-4557-8c80-28d8a58382d7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.311963] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.312209] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dac90a2-68c8-4e30-9f4a-69c2bc6d8598 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.317872] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Created folder: Project (b8efcd3690254590a3f636bba93809a6) in parent group-v268956. [ 1091.318058] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Creating folder: Instances. Parent ref: group-v269112. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1091.318280] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e16f669-2d43-41c7-8f6d-aeaf26f1a3e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.320727] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1091.320727] env[63175]: value = "task-1248657" [ 1091.320727] env[63175]: _type = "Task" [ 1091.320727] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.325601] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: cbcc2b72-e5c7-40b0-b51a-b3b49c3e3ad3] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1091.330275] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.332024] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Created folder: Instances in parent group-v269112. [ 1091.332374] env[63175]: DEBUG oslo.service.loopingcall [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.332600] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1091.332811] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5199c4a7-e036-4680-a93c-e272f99dbb43 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.349595] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Successfully created port: ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1091.354226] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1091.354226] env[63175]: value = "task-1248659" [ 1091.354226] env[63175]: _type = "Task" [ 1091.354226] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.362498] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248659, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.441636] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.442022] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.442388] env[63175]: DEBUG nova.objects.instance [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lazy-loading 'resources' on Instance uuid 86fd8d46-fcec-4d07-be79-d307d1fedcf0 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.564972] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1091.569110] env[63175]: DEBUG nova.compute.manager [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Received event network-changed-01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1091.569311] env[63175]: DEBUG nova.compute.manager [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Refreshing instance network info cache due to event network-changed-01dd3e04-4265-4a89-b499-1e99878db549. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1091.569526] env[63175]: DEBUG oslo_concurrency.lockutils [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] Acquiring lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.569674] env[63175]: DEBUG oslo_concurrency.lockutils [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] Acquired lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.569858] env[63175]: DEBUG nova.network.neutron [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Refreshing network info cache for port 01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1091.830987] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248657, 'name': PowerOffVM_Task, 'duration_secs': 0.20795} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.831601] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: f6d2c297-1cfe-4cab-8854-eb5760e6cbc0] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1091.834071] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.834071] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.834414] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c9637fd-b83c-45ce-8b6b-2df730464402 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.865779] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248659, 'name': CreateVM_Task, 'duration_secs': 0.436655} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.865779] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1091.865958] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.866019] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.866533] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1091.866631] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f78a83cf-a3d0-4739-b65b-4e5259bae725 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.872033] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1091.872033] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa275d-cb04-5660-24e7-964b45a6a577" [ 1091.872033] env[63175]: _type = "Task" [ 1091.872033] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.879326] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa275d-cb04-5660-24e7-964b45a6a577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.897082] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.897375] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.897574] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleting the datastore file [datastore2] 1579e4ad-f754-46ea-a9af-c85947489dcd {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.897844] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eda7e769-3cde-4647-97ec-4e482d9fd659 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.906024] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for the task: (returnval){ [ 1091.906024] env[63175]: value = "task-1248661" [ 1091.906024] env[63175]: _type = "Task" [ 1091.906024] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.912591] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.053032] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a8b2bd-3bb3-4c86-923d-b76b7f47a8e0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.060399] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db59ce04-506a-414c-840d-14c6326a8df7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.097027] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c16284-25ee-49d9-b197-c3f95588628f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.106310] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef179c26-abaf-459d-a376-b937e6590b4c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.119886] env[63175]: DEBUG nova.compute.provider_tree [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.301031] env[63175]: DEBUG nova.network.neutron [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Updated VIF entry in instance network info cache for port 01dd3e04-4265-4a89-b499-1e99878db549. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1092.301406] env[63175]: DEBUG nova.network.neutron [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Updating instance_info_cache with network_info: [{"id": "01dd3e04-4265-4a89-b499-1e99878db549", "address": "fa:16:3e:26:92:8c", "network": {"id": "36e1781f-25ae-41d0-bb2d-66ce72d0d672", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-937219789-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8efcd3690254590a3f636bba93809a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04ccbc7a-cf8d-4ea2-8411-291a1e27df7b", "external-id": "nsx-vlan-transportzone-998", "segmentation_id": 998, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01dd3e04-42", "ovs_interfaceid": "01dd3e04-4265-4a89-b499-1e99878db549", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.335036] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a2e70590-17ed-4804-b232-57526e87d22b] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1092.381964] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52aa275d-cb04-5660-24e7-964b45a6a577, 'name': SearchDatastore_Task, 'duration_secs': 0.008668} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.382292] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.382534] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1092.382774] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.382929] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.383128] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1092.383400] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf288785-8823-4049-a1e7-595baa674e45 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.397330] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1092.397546] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1092.398310] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d5175d3-dd33-40f7-a353-8928a92855be {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.403395] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1092.403395] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d502c6-7ffd-370d-6994-585675a5a914" [ 1092.403395] env[63175]: _type = "Task" [ 1092.403395] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.413972] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d502c6-7ffd-370d-6994-585675a5a914, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.416942] env[63175]: DEBUG oslo_vmware.api [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Task: {'id': task-1248661, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138713} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.417184] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.417371] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.417549] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.417758] env[63175]: INFO nova.compute.manager [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1092.418039] env[63175]: DEBUG oslo.service.loopingcall [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.418243] env[63175]: DEBUG nova.compute.manager [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1092.418339] env[63175]: DEBUG nova.network.neutron [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1092.598644] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1092.623382] env[63175]: DEBUG nova.scheduler.client.report [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1092.629117] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1092.629347] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1092.629514] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.629707] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1092.629863] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.630026] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1092.630290] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1092.630467] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1092.630642] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1092.630816] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1092.630991] env[63175]: DEBUG nova.virt.hardware [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.631853] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410b1cce-976e-4189-bb4d-90f6c32f6d05 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.640901] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b87437-cb79-4cfa-80f0-20f25488f294 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.804058] env[63175]: DEBUG oslo_concurrency.lockutils [req-5a7b6cbc-c1a7-4e0c-b8cc-d5919e0a57b3 req-05b3da08-ead9-4c71-a7b6-ded2f8174922 service nova] Releasing lock "refresh_cache-225096ef-500f-490f-b3c0-def1c46ed17c" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.808069] env[63175]: DEBUG nova.compute.manager [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Received event network-vif-plugged-ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1092.808289] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] Acquiring lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.808500] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.808669] env[63175]: DEBUG oslo_concurrency.lockutils [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.808889] env[63175]: DEBUG nova.compute.manager [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] No waiting events found dispatching network-vif-plugged-ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1092.809123] env[63175]: WARNING nova.compute.manager [req-b6daf870-5b66-4195-ad59-314b346a4d7e req-17c16ce5-4dd6-4af4-805f-4da6d9d9444a service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Received unexpected event network-vif-plugged-ec4dea97-52b6-435e-8ff3-52ea1cfe5614 for instance with vm_state building and task_state spawning. [ 1092.838081] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 35f775ef-8a9e-4c49-99b9-a90f8c8f39b4] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1092.913919] env[63175]: DEBUG nova.compute.manager [req-d06b6b4d-4f51-4fb5-837a-b6b099a30dd0 req-4af21b7e-37fd-4597-b439-a43d926dd729 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Received event network-vif-deleted-ebdfc338-744b-4ffc-8599-925b787e3901 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1092.914195] env[63175]: INFO nova.compute.manager [req-d06b6b4d-4f51-4fb5-837a-b6b099a30dd0 req-4af21b7e-37fd-4597-b439-a43d926dd729 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Neutron deleted interface ebdfc338-744b-4ffc-8599-925b787e3901; detaching it from the instance and deleting it from the info cache [ 1092.914378] env[63175]: DEBUG nova.network.neutron [req-d06b6b4d-4f51-4fb5-837a-b6b099a30dd0 req-4af21b7e-37fd-4597-b439-a43d926dd729 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.921768] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52d502c6-7ffd-370d-6994-585675a5a914, 'name': SearchDatastore_Task, 'duration_secs': 0.027762} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.922545] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4571cf7b-d73e-4cfc-971b-cc5a3130c6a4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.928063] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1092.928063] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0542c-d8b3-8415-8b8c-88753ded1cfb" [ 1092.928063] env[63175]: _type = "Task" [ 1092.928063] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.938227] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0542c-d8b3-8415-8b8c-88753ded1cfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.136037] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.160107] env[63175]: INFO nova.scheduler.client.report [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Deleted allocations for instance 86fd8d46-fcec-4d07-be79-d307d1fedcf0 [ 1093.342056] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: c445245b-b7d3-49c6-82c5-1e8188c89b68] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1093.362498] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Successfully updated port: ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1093.384752] env[63175]: DEBUG nova.network.neutron [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.418394] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6656c4b-7845-4a8d-96a0-d0e5f96394b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.428088] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ec11ef-ed80-475d-9a61-26861216129a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.448390] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52c0542c-d8b3-8415-8b8c-88753ded1cfb, 'name': SearchDatastore_Task, 'duration_secs': 0.01122} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.448669] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.448937] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 225096ef-500f-490f-b3c0-def1c46ed17c/225096ef-500f-490f-b3c0-def1c46ed17c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1093.449217] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d14eb363-3826-4972-b1c0-7171ea3be152 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.456395] env[63175]: DEBUG nova.compute.manager [req-d06b6b4d-4f51-4fb5-837a-b6b099a30dd0 req-4af21b7e-37fd-4597-b439-a43d926dd729 service nova] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Detach interface failed, port_id=ebdfc338-744b-4ffc-8599-925b787e3901, reason: Instance 1579e4ad-f754-46ea-a9af-c85947489dcd could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1093.461522] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1093.461522] env[63175]: value = "task-1248662" [ 1093.461522] env[63175]: _type = "Task" [ 1093.461522] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.471129] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.597480] env[63175]: DEBUG nova.compute.manager [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Received event network-changed-ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1093.597735] env[63175]: DEBUG nova.compute.manager [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Refreshing instance network info cache due to event network-changed-ec4dea97-52b6-435e-8ff3-52ea1cfe5614. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1093.598084] env[63175]: DEBUG oslo_concurrency.lockutils [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] Acquiring lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.598273] env[63175]: DEBUG oslo_concurrency.lockutils [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] Acquired lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.598472] env[63175]: DEBUG nova.network.neutron [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Refreshing network info cache for port ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1093.667510] env[63175]: DEBUG oslo_concurrency.lockutils [None req-12598fa6-ac5a-4e12-b9e6-0a2d65f55181 tempest-ServerShowV257Test-1586487661 tempest-ServerShowV257Test-1586487661-project-member] Lock "86fd8d46-fcec-4d07-be79-d307d1fedcf0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.411s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.844662] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 39e55060-73a7-4dbd-96cf-bc48d8737c1c] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1093.866278] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.887453] env[63175]: INFO nova.compute.manager [-] [instance: 1579e4ad-f754-46ea-a9af-c85947489dcd] Took 1.47 seconds to deallocate network for instance. [ 1093.971685] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44604} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.971996] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 225096ef-500f-490f-b3c0-def1c46ed17c/225096ef-500f-490f-b3c0-def1c46ed17c.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1093.972258] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1093.972537] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f065aa94-25fc-4def-8896-8243748d691a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.979049] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1093.979049] env[63175]: value = "task-1248663" [ 1093.979049] env[63175]: _type = "Task" [ 1093.979049] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.986891] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.128412] env[63175]: DEBUG nova.network.neutron [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1094.204226] env[63175]: DEBUG nova.network.neutron [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.348404] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 36089589-d105-49e0-8ae7-790c814b036c] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1094.393691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.393959] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.394254] env[63175]: DEBUG nova.objects.instance [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lazy-loading 'resources' on Instance uuid 1579e4ad-f754-46ea-a9af-c85947489dcd {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.489548] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072357} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.489823] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1094.490614] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9421baa3-4d8f-4562-9630-85cf289395d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.513044] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 225096ef-500f-490f-b3c0-def1c46ed17c/225096ef-500f-490f-b3c0-def1c46ed17c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.513310] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01a506e2-5b4a-430e-94dc-f1d180857f01 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.532045] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1094.532045] env[63175]: value = "task-1248664" [ 1094.532045] env[63175]: _type = "Task" [ 1094.532045] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.539808] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248664, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.706359] env[63175]: DEBUG oslo_concurrency.lockutils [req-4cd8992f-b84b-45ce-832a-03a04a57af35 req-6df38805-e49e-497a-92bf-2a7438497b13 service nova] Releasing lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.706768] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.706928] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.851067] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 67e8715d-b729-4013-8cca-44eaa55c662e] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1094.967138] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1daa3c-fcef-4ae8-8ed5-c0d95298043e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.974493] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b113e61e-a098-464d-ac21-f559424ac4a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.005023] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc7b383-86b6-4279-8817-35094a19ea9a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.011914] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44496ea3-2618-4a92-bb92-98c84c4cf15e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.024524] env[63175]: DEBUG nova.compute.provider_tree [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.040570] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248664, 'name': ReconfigVM_Task, 'duration_secs': 0.301724} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.040895] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 225096ef-500f-490f-b3c0-def1c46ed17c/225096ef-500f-490f-b3c0-def1c46ed17c.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.041410] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-143261bc-d747-4044-a2e3-68a21e55ba6f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.046695] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1095.046695] env[63175]: value = "task-1248665" [ 1095.046695] env[63175]: _type = "Task" [ 1095.046695] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.053870] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248665, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.236885] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1095.354507] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 0f8e580e-fb14-4db8-b995-a9ffe06d8bac] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1095.365142] env[63175]: DEBUG nova.network.neutron [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Updating instance_info_cache with network_info: [{"id": "ec4dea97-52b6-435e-8ff3-52ea1cfe5614", "address": "fa:16:3e:1f:1d:e9", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4dea97-52", "ovs_interfaceid": "ec4dea97-52b6-435e-8ff3-52ea1cfe5614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.528013] env[63175]: DEBUG nova.scheduler.client.report [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1095.556779] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248665, 'name': Rename_Task} progress is 14%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.857594] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: e137904f-dc43-4ebb-90ab-e10ea5487fe5] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1095.867541] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-00eb04f9-d250-4f18-8f88-16dbcc2c2958" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.867844] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Instance network_info: |[{"id": "ec4dea97-52b6-435e-8ff3-52ea1cfe5614", "address": "fa:16:3e:1f:1d:e9", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4dea97-52", "ovs_interfaceid": "ec4dea97-52b6-435e-8ff3-52ea1cfe5614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1095.868314] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:1d:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec4dea97-52b6-435e-8ff3-52ea1cfe5614', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.876263] env[63175]: DEBUG oslo.service.loopingcall [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.876492] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1095.877250] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ba0d558-0143-4e6c-b8f9-751743a79239 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.897489] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.897489] env[63175]: value = "task-1248666" [ 1095.897489] env[63175]: _type = "Task" [ 1095.897489] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.905520] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248666, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.033372] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.051844] env[63175]: INFO nova.scheduler.client.report [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Deleted allocations for instance 1579e4ad-f754-46ea-a9af-c85947489dcd [ 1096.060343] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248665, 'name': Rename_Task, 'duration_secs': 0.840835} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.062866] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1096.063327] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0d5c9a1-4a6f-4660-882f-2508890fa1a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.070227] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1096.070227] env[63175]: value = "task-1248667" [ 1096.070227] env[63175]: _type = "Task" [ 1096.070227] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.080356] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.361334] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 12a2d788-c995-468b-b2a2-17dba8ad01db] Instance has had 0 of 5 cleanup attempts {{(pid=63175) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1096.408883] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248666, 'name': CreateVM_Task, 'duration_secs': 0.308552} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.409072] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1096.409831] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.410038] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.410364] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.410629] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5082283a-9d0b-4b58-a38a-dee8a1605a6a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.415841] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1096.415841] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893b18-9f03-18fe-948e-789e15d47cb9" [ 1096.415841] env[63175]: _type = "Task" [ 1096.415841] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.424171] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893b18-9f03-18fe-948e-789e15d47cb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.567084] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f0f201fe-030f-43d4-be36-be16c46a9baf tempest-AttachVolumeNegativeTest-1816329101 tempest-AttachVolumeNegativeTest-1816329101-project-member] Lock "1579e4ad-f754-46ea-a9af-c85947489dcd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.772s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.584240] env[63175]: DEBUG oslo_vmware.api [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248667, 'name': PowerOnVM_Task, 'duration_secs': 0.446435} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.585379] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1096.585748] env[63175]: INFO nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Took 7.20 seconds to spawn the instance on the hypervisor. [ 1096.586044] env[63175]: DEBUG nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1096.587390] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df45628-48f5-47ca-9926-2e96c357ee90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.865554] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1096.865554] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Cleaning up deleted instances with incomplete migration {{(pid=63175) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 1096.927298] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52893b18-9f03-18fe-948e-789e15d47cb9, 'name': SearchDatastore_Task, 'duration_secs': 0.009021} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.927610] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.927888] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.928163] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.928319] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.928508] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.928832] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23fe6734-6d90-4a18-97b9-7774dfc28037 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.941329] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.941736] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1096.942730] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e178f575-02ef-485c-af3b-221dddb36bac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.951848] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1096.951848] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5220273a-0e65-b312-c9b0-34ee78df4439" [ 1096.951848] env[63175]: _type = "Task" [ 1096.951848] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.962584] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5220273a-0e65-b312-c9b0-34ee78df4439, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.109236] env[63175]: INFO nova.compute.manager [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Took 12.01 seconds to build instance. [ 1097.368061] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.463737] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5220273a-0e65-b312-c9b0-34ee78df4439, 'name': SearchDatastore_Task, 'duration_secs': 0.01074} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.464612] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52f93bdb-2163-4436-9cf5-ab621f042ac1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.471086] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1097.471086] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fa9556-2383-ae94-d213-671b8d193ba0" [ 1097.471086] env[63175]: _type = "Task" [ 1097.471086] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.481273] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fa9556-2383-ae94-d213-671b8d193ba0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.610841] env[63175]: DEBUG oslo_concurrency.lockutils [None req-c0dfcbf7-0e36-4a61-879d-aa92547661b8 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.521s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.981936] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52fa9556-2383-ae94-d213-671b8d193ba0, 'name': SearchDatastore_Task, 'duration_secs': 0.009753} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.982344] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.982512] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 00eb04f9-d250-4f18-8f88-16dbcc2c2958/00eb04f9-d250-4f18-8f88-16dbcc2c2958.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1097.982782] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f68e5e4f-2816-47f0-8c04-261424cf5ec6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.991140] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1097.991140] env[63175]: value = "task-1248669" [ 1097.991140] env[63175]: _type = "Task" [ 1097.991140] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.000620] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248669, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.501699] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248669, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.795992] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "225096ef-500f-490f-b3c0-def1c46ed17c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.796233] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.796461] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.796648] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.796855] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.799146] env[63175]: INFO nova.compute.manager [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Terminating instance [ 1099.001673] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248669, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523609} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.002031] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 00eb04f9-d250-4f18-8f88-16dbcc2c2958/00eb04f9-d250-4f18-8f88-16dbcc2c2958.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1099.002226] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1099.002482] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-750f44a9-f6ed-401b-af5e-4f2837e23438 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.009283] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1099.009283] env[63175]: value = "task-1248670" [ 1099.009283] env[63175]: _type = "Task" [ 1099.009283] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.017790] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248670, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.303706] env[63175]: DEBUG nova.compute.manager [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1099.304127] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1099.305183] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a7e716-0f5c-4f95-95f3-06961b31f6cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.313801] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.314091] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9654148f-c006-4dcc-8cce-47d40946d1aa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.321222] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1099.321222] env[63175]: value = "task-1248671" [ 1099.321222] env[63175]: _type = "Task" [ 1099.321222] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.330690] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.519808] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248670, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235368} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.520126] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.521093] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7a36c3-d61f-4d42-b2aa-85e9a97ba31b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.543790] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 00eb04f9-d250-4f18-8f88-16dbcc2c2958/00eb04f9-d250-4f18-8f88-16dbcc2c2958.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.544290] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-409f6ae6-caa4-4c34-82c0-d63f72c370e5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.563661] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1099.563661] env[63175]: value = "task-1248673" [ 1099.563661] env[63175]: _type = "Task" [ 1099.563661] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.571869] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.832119] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248671, 'name': PowerOffVM_Task, 'duration_secs': 0.230795} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.832397] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.832608] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.832858] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfaf6381-8939-4515-b3bf-9a985bf9b783 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.899494] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.899744] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.899972] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Deleting the datastore file [datastore2] 225096ef-500f-490f-b3c0-def1c46ed17c {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.900425] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75f6d0f3-cd7f-4a37-974a-8d268f7c24a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.906348] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for the task: (returnval){ [ 1099.906348] env[63175]: value = "task-1248675" [ 1099.906348] env[63175]: _type = "Task" [ 1099.906348] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.914457] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248675, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.073745] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248673, 'name': ReconfigVM_Task, 'duration_secs': 0.289401} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.074130] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 00eb04f9-d250-4f18-8f88-16dbcc2c2958/00eb04f9-d250-4f18-8f88-16dbcc2c2958.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.074655] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ce1c815-ebc2-4472-b04b-35cb595126d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.081345] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1100.081345] env[63175]: value = "task-1248676" [ 1100.081345] env[63175]: _type = "Task" [ 1100.081345] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.088516] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248676, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.416325] env[63175]: DEBUG oslo_vmware.api [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Task: {'id': task-1248675, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245541} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.416582] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.416818] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.417020] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.417204] env[63175]: INFO nova.compute.manager [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1100.417442] env[63175]: DEBUG oslo.service.loopingcall [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.417641] env[63175]: DEBUG nova.compute.manager [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1100.417738] env[63175]: DEBUG nova.network.neutron [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1100.591722] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248676, 'name': Rename_Task, 'duration_secs': 0.159775} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.592080] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.592277] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53321783-5e96-40cf-9ab1-fc50af5370fd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.598401] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1100.598401] env[63175]: value = "task-1248677" [ 1100.598401] env[63175]: _type = "Task" [ 1100.598401] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.605901] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.660066] env[63175]: DEBUG nova.compute.manager [req-90c314d5-258e-4b34-bbbd-49d165bc254d req-985d78a0-5e40-449a-bfd2-b6194d5dd9d7 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Received event network-vif-deleted-01dd3e04-4265-4a89-b499-1e99878db549 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1100.660323] env[63175]: INFO nova.compute.manager [req-90c314d5-258e-4b34-bbbd-49d165bc254d req-985d78a0-5e40-449a-bfd2-b6194d5dd9d7 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Neutron deleted interface 01dd3e04-4265-4a89-b499-1e99878db549; detaching it from the instance and deleting it from the info cache [ 1100.660999] env[63175]: DEBUG nova.network.neutron [req-90c314d5-258e-4b34-bbbd-49d165bc254d req-985d78a0-5e40-449a-bfd2-b6194d5dd9d7 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.870098] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.870307] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.107996] env[63175]: DEBUG oslo_vmware.api [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248677, 'name': PowerOnVM_Task, 'duration_secs': 0.504333} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.108374] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.108526] env[63175]: INFO nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1101.108709] env[63175]: DEBUG nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1101.109478] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba194fa-7335-4131-aaaf-b1d87e7d8ec8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.134906] env[63175]: DEBUG nova.network.neutron [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.163662] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0cd2cdd1-3cbe-4080-a878-da1030694091 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.173935] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31408c12-e0a0-4c68-a0ca-5f5b374dd804 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.203669] env[63175]: DEBUG nova.compute.manager [req-90c314d5-258e-4b34-bbbd-49d165bc254d req-985d78a0-5e40-449a-bfd2-b6194d5dd9d7 service nova] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Detach interface failed, port_id=01dd3e04-4265-4a89-b499-1e99878db549, reason: Instance 225096ef-500f-490f-b3c0-def1c46ed17c could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1101.375922] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.375922] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1101.376130] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1101.627236] env[63175]: INFO nova.compute.manager [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Took 14.53 seconds to build instance. [ 1101.637935] env[63175]: INFO nova.compute.manager [-] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Took 1.22 seconds to deallocate network for instance. [ 1101.879217] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 225096ef-500f-490f-b3c0-def1c46ed17c] Skipping network cache update for instance because it is being deleted. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10325}} [ 1101.906075] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.906237] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquired lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.906377] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Forcefully refreshing network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1101.906542] env[63175]: DEBUG nova.objects.instance [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lazy-loading 'info_cache' on Instance uuid 5d4d277e-ec2d-40ed-b16a-6d87965888e2 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.129633] env[63175]: DEBUG oslo_concurrency.lockutils [None req-2efe175f-2cc1-466b-be81-3196d7fab637 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.044s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.143832] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.144140] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.144373] env[63175]: DEBUG nova.objects.instance [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lazy-loading 'resources' on Instance uuid 225096ef-500f-490f-b3c0-def1c46ed17c {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.746820] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.749595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.749595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.749595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.749595] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.751298] env[63175]: INFO nova.compute.manager [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Terminating instance [ 1102.753959] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e208a5f0-4d50-4edf-a541-fb408a6f6cee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.770331] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ba1f27-7fcb-4829-94f9-0e0b75ade89a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.804476] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa98f98-9613-47a9-9873-ce0ed2582c0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.811646] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f62e86d-76f6-426e-b003-5793d7bc73bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.824989] env[63175]: DEBUG nova.compute.provider_tree [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.260699] env[63175]: DEBUG nova.compute.manager [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1103.260986] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1103.261831] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2d4011-0d3d-403a-bb32-1571bc60c81b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.269904] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.270147] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6de64012-8308-43bc-991c-560dd88c6cd8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.276426] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1103.276426] env[63175]: value = "task-1248679" [ 1103.276426] env[63175]: _type = "Task" [ 1103.276426] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.283596] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.327836] env[63175]: DEBUG nova.scheduler.client.report [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1103.619800] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updating instance_info_cache with network_info: [{"id": "72df71c2-faab-4cc4-b2c3-e730bb625453", "address": "fa:16:3e:f3:ef:b6", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72df71c2-fa", "ovs_interfaceid": "72df71c2-faab-4cc4-b2c3-e730bb625453", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.786631] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248679, 'name': PowerOffVM_Task, 'duration_secs': 0.214446} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.786631] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.786779] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1103.786940] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb9ede9d-f89c-46a1-b43d-b67116dcf939 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.832424] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.688s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.850265] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.850512] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.850704] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore2] 00eb04f9-d250-4f18-8f88-16dbcc2c2958 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.851686] env[63175]: INFO nova.scheduler.client.report [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Deleted allocations for instance 225096ef-500f-490f-b3c0-def1c46ed17c [ 1103.852658] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9235c5e2-9f5b-468f-8d6f-58f461534e58 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.862739] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1103.862739] env[63175]: value = "task-1248681" [ 1103.862739] env[63175]: _type = "Task" [ 1103.862739] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.869871] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.122182] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Releasing lock "refresh_cache-5d4d277e-ec2d-40ed-b16a-6d87965888e2" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.122378] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updated the network info_cache for instance {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1104.122559] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.122725] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.122877] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.123038] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.123188] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.123342] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_power_states {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.362298] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24900e2f-25bb-4529-b1df-092b792a3344 tempest-ServersNegativeTestMultiTenantJSON-1676010416 tempest-ServersNegativeTestMultiTenantJSON-1676010416-project-member] Lock "225096ef-500f-490f-b3c0-def1c46ed17c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.566s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.372231] env[63175]: DEBUG oslo_vmware.api [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125702} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.372449] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.372633] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1104.372835] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1104.372998] env[63175]: INFO nova.compute.manager [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1104.373250] env[63175]: DEBUG oslo.service.loopingcall [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.373446] env[63175]: DEBUG nova.compute.manager [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1104.373580] env[63175]: DEBUG nova.network.neutron [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1104.627140] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Getting list of instances from cluster (obj){ [ 1104.627140] env[63175]: value = "domain-c8" [ 1104.627140] env[63175]: _type = "ClusterComputeResource" [ 1104.627140] env[63175]: } {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1104.628196] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf19aadf-b146-440a-b60c-9840c1ad48a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.637764] env[63175]: DEBUG nova.compute.manager [req-eab34dc9-0bd3-4d81-b218-a296aa67f03f req-f4f063b0-f223-46f9-b86f-da59f5346fe4 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Received event network-vif-deleted-ec4dea97-52b6-435e-8ff3-52ea1cfe5614 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1104.637972] env[63175]: INFO nova.compute.manager [req-eab34dc9-0bd3-4d81-b218-a296aa67f03f req-f4f063b0-f223-46f9-b86f-da59f5346fe4 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Neutron deleted interface ec4dea97-52b6-435e-8ff3-52ea1cfe5614; detaching it from the instance and deleting it from the info cache [ 1104.638174] env[63175]: DEBUG nova.network.neutron [req-eab34dc9-0bd3-4d81-b218-a296aa67f03f req-f4f063b0-f223-46f9-b86f-da59f5346fe4 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.646663] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Got total of 2 instances {{(pid=63175) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1104.646918] env[63175]: WARNING nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1104.647055] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Triggering sync for uuid 5d4d277e-ec2d-40ed-b16a-6d87965888e2 {{(pid=63175) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1104.647249] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Triggering sync for uuid 45a4df18-72bf-40c5-a351-827221dd44db {{(pid=63175) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1104.647399] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Triggering sync for uuid 00eb04f9-d250-4f18-8f88-16dbcc2c2958 {{(pid=63175) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1104.647703] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.647900] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.648197] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "45a4df18-72bf-40c5-a351-827221dd44db" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.648382] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "45a4df18-72bf-40c5-a351-827221dd44db" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.649015] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.649015] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.649015] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1104.649648] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50271c8a-7192-42bb-a756-94bad7cd00ff {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.653052] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df0abb7-a08c-4f7a-9b09-54fb04c2d348 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.655517] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.104259] env[63175]: DEBUG nova.network.neutron [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.141403] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8c102d1-ce66-48ba-9997-c40f6d60977c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.151494] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6a2f51-2519-4390-83ad-965408a836dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.162949] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.163192] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.163355] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.163503] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1105.164370] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fb7edc-cf78-42bd-a55a-586afb90284c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.167899] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "45a4df18-72bf-40c5-a351-827221dd44db" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.519s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.168388] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.174026] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd90cd7-a2f5-418c-80db-ff7039c5b328 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.183708] env[63175]: DEBUG nova.compute.manager [req-eab34dc9-0bd3-4d81-b218-a296aa67f03f req-f4f063b0-f223-46f9-b86f-da59f5346fe4 service nova] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Detach interface failed, port_id=ec4dea97-52b6-435e-8ff3-52ea1cfe5614, reason: Instance 00eb04f9-d250-4f18-8f88-16dbcc2c2958 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1105.196401] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e62120c-a280-4701-a4c8-7b43341b615e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.203238] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a05349-3d53-43ea-ba9c-0f156f02ff0a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.234482] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180441MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1105.234643] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.234873] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.606847] env[63175]: INFO nova.compute.manager [-] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] Took 1.23 seconds to deallocate network for instance. [ 1106.113773] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.263974] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 5d4d277e-ec2d-40ed-b16a-6d87965888e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1106.264158] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 45a4df18-72bf-40c5-a351-827221dd44db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1106.264282] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance 00eb04f9-d250-4f18-8f88-16dbcc2c2958 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1106.264475] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1106.264613] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1106.325275] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df93b16-17c3-4c3c-bb77-71376e5fb91b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.333551] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d3f281-1786-4dcd-ba63-385d7801274f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.364091] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060d97ed-9a86-4439-b009-669d8a74557c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.371768] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b385e9-5637-475a-a268-ccb51ca9eec1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.386501] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.892032] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1107.395153] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1107.395378] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.160s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.395667] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.282s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.395937] env[63175]: DEBUG nova.objects.instance [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 00eb04f9-d250-4f18-8f88-16dbcc2c2958 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.952574] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da3c004-0bd0-46bb-944d-8a56c7ff59c5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.960625] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fb23bb-a49e-4ecf-8568-a1db4831463f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.990369] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740aceb7-fc63-4437-adba-898db66ea795 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.997345] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4b5ec8-1aa4-48a4-80af-f86129bb8b70 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.009879] env[63175]: DEBUG nova.compute.provider_tree [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.513050] env[63175]: DEBUG nova.scheduler.client.report [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1109.017756] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.042124] env[63175]: INFO nova.scheduler.client.report [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 00eb04f9-d250-4f18-8f88-16dbcc2c2958 [ 1109.550983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-f066a86c-7cc9-4029-9cea-2db83c1875f9 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.803s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.552291] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.904s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.552491] env[63175]: INFO nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: 00eb04f9-d250-4f18-8f88-16dbcc2c2958] During sync_power_state the instance has a pending task (deleting). Skip. [ 1109.552673] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "00eb04f9-d250-4f18-8f88-16dbcc2c2958" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.469203] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "45a4df18-72bf-40c5-a351-827221dd44db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.469556] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.469691] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "45a4df18-72bf-40c5-a351-827221dd44db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.469879] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.470064] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.472198] env[63175]: INFO nova.compute.manager [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Terminating instance [ 1110.976228] env[63175]: DEBUG nova.compute.manager [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1110.976508] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1110.977506] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f89088b-f70a-4bcf-8eef-8d5ae3356a76 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.985451] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.985682] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4d35a04-b0cc-440f-8ef6-b787ab9ca1f0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.992499] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1110.992499] env[63175]: value = "task-1248682" [ 1110.992499] env[63175]: _type = "Task" [ 1110.992499] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.000266] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.502027] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248682, 'name': PowerOffVM_Task, 'duration_secs': 0.201184} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.502368] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.502486] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.502714] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d7192f9-df9a-42a1-ac77-15be379100d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.572904] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.573151] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.573339] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore1] 45a4df18-72bf-40c5-a351-827221dd44db {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.573607] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ecad03b-1eb4-4d56-b486-3d73c5e6515f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.579766] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1111.579766] env[63175]: value = "task-1248684" [ 1111.579766] env[63175]: _type = "Task" [ 1111.579766] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.587361] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.089768] env[63175]: DEBUG oslo_vmware.api [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172378} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.089990] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.090169] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.090358] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.090537] env[63175]: INFO nova.compute.manager [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1112.090779] env[63175]: DEBUG oslo.service.loopingcall [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.090973] env[63175]: DEBUG nova.compute.manager [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1112.091081] env[63175]: DEBUG nova.network.neutron [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1112.324118] env[63175]: DEBUG nova.compute.manager [req-90d0596e-27d1-40c6-890c-6ae3fb21ede3 req-aa3bf97d-a692-4930-b87a-2712ecea81fb service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Received event network-vif-deleted-8ea56d43-c9b6-45a9-aaf8-98da64c6ee80 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1112.324118] env[63175]: INFO nova.compute.manager [req-90d0596e-27d1-40c6-890c-6ae3fb21ede3 req-aa3bf97d-a692-4930-b87a-2712ecea81fb service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Neutron deleted interface 8ea56d43-c9b6-45a9-aaf8-98da64c6ee80; detaching it from the instance and deleting it from the info cache [ 1112.324118] env[63175]: DEBUG nova.network.neutron [req-90d0596e-27d1-40c6-890c-6ae3fb21ede3 req-aa3bf97d-a692-4930-b87a-2712ecea81fb service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.802376] env[63175]: DEBUG nova.network.neutron [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.825457] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-001d3f2c-3631-4d76-8d26-fa9427de8333 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.835530] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8c4d24-cd6b-4e35-bdc1-c5d9014b5c58 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.859549] env[63175]: DEBUG nova.compute.manager [req-90d0596e-27d1-40c6-890c-6ae3fb21ede3 req-aa3bf97d-a692-4930-b87a-2712ecea81fb service nova] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Detach interface failed, port_id=8ea56d43-c9b6-45a9-aaf8-98da64c6ee80, reason: Instance 45a4df18-72bf-40c5-a351-827221dd44db could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1113.305111] env[63175]: INFO nova.compute.manager [-] [instance: 45a4df18-72bf-40c5-a351-827221dd44db] Took 1.21 seconds to deallocate network for instance. [ 1113.812216] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.812557] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.812730] env[63175]: DEBUG nova.objects.instance [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 45a4df18-72bf-40c5-a351-827221dd44db {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.356694] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06153564-68a8-4132-bc1d-5d25bf12b8cd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.363877] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7ea85c-d2ab-41ec-9829-7ff41e9ee4ab {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.393131] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fc1321-5180-49b3-9005-891470e933e0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.400134] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d1dd27-08d1-47e7-bc5e-8c9a9b21cc13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.412708] env[63175]: DEBUG nova.compute.provider_tree [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.916029] env[63175]: DEBUG nova.scheduler.client.report [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1115.420892] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.438424] env[63175]: INFO nova.scheduler.client.report [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 45a4df18-72bf-40c5-a351-827221dd44db [ 1115.949258] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3da53416-9d7f-4088-bc9e-c1aa0dd6a004 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "45a4df18-72bf-40c5-a351-827221dd44db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.480s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.472321] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d676662-2185-4ef2-b9ab-b4b907992566" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.472647] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.975018] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1118.494967] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.495264] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.496691] env[63175]: INFO nova.compute.claims [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.540195] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3921c1-708d-4460-8e97-1f06f8f65ca7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.547805] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b4ff11-26a9-4886-b73a-5a27079ca1f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.578331] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267b1e9c-800b-468f-b649-fb7694e502fb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.585245] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5cf363-2b03-478c-b446-44eb55f5daa1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.599754] env[63175]: DEBUG nova.compute.provider_tree [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.103056] env[63175]: DEBUG nova.scheduler.client.report [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1120.607201] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.112s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.607770] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1121.112439] env[63175]: DEBUG nova.compute.utils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.113859] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1121.114040] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1121.161350] env[63175]: DEBUG nova.policy [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1121.401600] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Successfully created port: e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1121.617278] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1122.627185] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1122.652783] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.653055] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.653231] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.653425] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.653581] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.653758] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.653955] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.654134] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.654312] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.654480] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.654655] env[63175]: DEBUG nova.virt.hardware [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.655587] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7184a6ab-2fb3-448e-a77b-aac3e6ccda2a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.663525] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d174ab42-3d4e-4cf8-83b7-b9bf42b68473 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.756038] env[63175]: DEBUG nova.compute.manager [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Received event network-vif-plugged-e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1122.756279] env[63175]: DEBUG oslo_concurrency.lockutils [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] Acquiring lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.756500] env[63175]: DEBUG oslo_concurrency.lockutils [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] Lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.756671] env[63175]: DEBUG oslo_concurrency.lockutils [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] Lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.756845] env[63175]: DEBUG nova.compute.manager [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] No waiting events found dispatching network-vif-plugged-e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1122.757037] env[63175]: WARNING nova.compute.manager [req-4c8d7e35-62be-4039-9efb-edd56d7f4d66 req-f48916f6-c144-4192-856d-cbfed01353d6 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Received unexpected event network-vif-plugged-e0d257f6-bdfe-42e0-a4c2-5d8425081a51 for instance with vm_state building and task_state spawning. [ 1122.840983] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Successfully updated port: e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1123.344704] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.344833] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.345016] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1123.897453] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1124.019116] env[63175]: DEBUG nova.network.neutron [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Updating instance_info_cache with network_info: [{"id": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "address": "fa:16:3e:31:21:64", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d257f6-bd", "ovs_interfaceid": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.521868] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.522202] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Instance network_info: |[{"id": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "address": "fa:16:3e:31:21:64", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d257f6-bd", "ovs_interfaceid": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1124.522653] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:21:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0d257f6-bdfe-42e0-a4c2-5d8425081a51', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.529942] env[63175]: DEBUG oslo.service.loopingcall [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.530165] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1124.530383] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3de22594-8e06-4e5d-8bde-e7c538b18c80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.550206] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.550206] env[63175]: value = "task-1248685" [ 1124.550206] env[63175]: _type = "Task" [ 1124.550206] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.557646] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248685, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.782225] env[63175]: DEBUG nova.compute.manager [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Received event network-changed-e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1124.782424] env[63175]: DEBUG nova.compute.manager [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Refreshing instance network info cache due to event network-changed-e0d257f6-bdfe-42e0-a4c2-5d8425081a51. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1124.782648] env[63175]: DEBUG oslo_concurrency.lockutils [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] Acquiring lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.782804] env[63175]: DEBUG oslo_concurrency.lockutils [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] Acquired lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.782975] env[63175]: DEBUG nova.network.neutron [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Refreshing network info cache for port e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1125.060462] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248685, 'name': CreateVM_Task, 'duration_secs': 0.300384} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.060743] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1125.061255] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.061422] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.061728] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.061967] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-706a913a-43e3-4d0b-b7ff-8d79422b9687 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.066447] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1125.066447] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b36cdc-9b2e-faae-321a-eadb9feee2ba" [ 1125.066447] env[63175]: _type = "Task" [ 1125.066447] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.073823] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b36cdc-9b2e-faae-321a-eadb9feee2ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.480424] env[63175]: DEBUG nova.network.neutron [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Updated VIF entry in instance network info cache for port e0d257f6-bdfe-42e0-a4c2-5d8425081a51. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1125.480793] env[63175]: DEBUG nova.network.neutron [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Updating instance_info_cache with network_info: [{"id": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "address": "fa:16:3e:31:21:64", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d257f6-bd", "ovs_interfaceid": "e0d257f6-bdfe-42e0-a4c2-5d8425081a51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.576464] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52b36cdc-9b2e-faae-321a-eadb9feee2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.010353} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.576749] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.577011] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.577270] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.577423] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.577607] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.577862] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38a7f653-3910-426f-8afd-8775bf650b73 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.585542] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.585672] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1125.586368] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbfdb029-4f7d-45e4-8770-5ae8b08a89a8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.591017] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1125.591017] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a770dc-b9eb-cc0c-9436-42a7d92b9417" [ 1125.591017] env[63175]: _type = "Task" [ 1125.591017] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.597866] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a770dc-b9eb-cc0c-9436-42a7d92b9417, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.983321] env[63175]: DEBUG oslo_concurrency.lockutils [req-b33885a2-7db6-433b-b036-f068a0930184 req-ce1d6495-d708-447d-a7b3-458f49695192 service nova] Releasing lock "refresh_cache-5d676662-2185-4ef2-b9ab-b4b907992566" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.101402] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a770dc-b9eb-cc0c-9436-42a7d92b9417, 'name': SearchDatastore_Task, 'duration_secs': 0.008009} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.102195] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c09eb7cc-3e04-42d5-8ed5-715065f40c27 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.107120] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1126.107120] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52646b34-900f-b479-f1dd-86d8360bf876" [ 1126.107120] env[63175]: _type = "Task" [ 1126.107120] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.114316] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52646b34-900f-b479-f1dd-86d8360bf876, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.618463] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52646b34-900f-b479-f1dd-86d8360bf876, 'name': SearchDatastore_Task, 'duration_secs': 0.008467} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.618716] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.618976] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5d676662-2185-4ef2-b9ab-b4b907992566/5d676662-2185-4ef2-b9ab-b4b907992566.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1126.619247] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70d8f83e-26aa-405b-91d3-1bc23a124b01 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.625865] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1126.625865] env[63175]: value = "task-1248686" [ 1126.625865] env[63175]: _type = "Task" [ 1126.625865] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.632922] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.136076] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.406682} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.136436] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore2] 5d676662-2185-4ef2-b9ab-b4b907992566/5d676662-2185-4ef2-b9ab-b4b907992566.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1127.136526] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.136776] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f8f8845-6597-466a-a5ef-168239ee0c4e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.143272] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1127.143272] env[63175]: value = "task-1248687" [ 1127.143272] env[63175]: _type = "Task" [ 1127.143272] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.150093] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.653113] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062459} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.653435] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.654175] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0f5b9f-2e90-4cec-8678-8565aa36d4d2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.675058] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 5d676662-2185-4ef2-b9ab-b4b907992566/5d676662-2185-4ef2-b9ab-b4b907992566.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.675299] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36db2476-58c1-406a-ab18-a2d71e02e748 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.693095] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1127.693095] env[63175]: value = "task-1248688" [ 1127.693095] env[63175]: _type = "Task" [ 1127.693095] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.699982] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.203925] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248688, 'name': ReconfigVM_Task, 'duration_secs': 0.269266} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.204363] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 5d676662-2185-4ef2-b9ab-b4b907992566/5d676662-2185-4ef2-b9ab-b4b907992566.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.204890] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61c0250b-653a-4b00-a8d4-a164a2370e22 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.211070] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1128.211070] env[63175]: value = "task-1248689" [ 1128.211070] env[63175]: _type = "Task" [ 1128.211070] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.218622] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248689, 'name': Rename_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.720353] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248689, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.221875] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248689, 'name': Rename_Task} progress is 99%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.722953] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248689, 'name': Rename_Task, 'duration_secs': 1.118605} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.723261] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1129.723511] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8770f68-7c24-45a1-99e5-a16953c21cd9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.729261] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1129.729261] env[63175]: value = "task-1248690" [ 1129.729261] env[63175]: _type = "Task" [ 1129.729261] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.736440] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.239385] env[63175]: DEBUG oslo_vmware.api [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248690, 'name': PowerOnVM_Task, 'duration_secs': 0.475159} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.239753] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1130.239869] env[63175]: INFO nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Took 7.61 seconds to spawn the instance on the hypervisor. [ 1130.240065] env[63175]: DEBUG nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1130.240865] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c18bc4-3e0f-48db-8132-3b2392529c9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.761362] env[63175]: INFO nova.compute.manager [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Took 12.28 seconds to build instance. [ 1131.262857] env[63175]: DEBUG oslo_concurrency.lockutils [None req-e967fb3f-b06f-4f4e-b11e-517b376d24e1 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.790s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.081359] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d676662-2185-4ef2-b9ab-b4b907992566" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.081627] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.081839] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.082039] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.082226] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.084450] env[63175]: INFO nova.compute.manager [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Terminating instance [ 1132.588572] env[63175]: DEBUG nova.compute.manager [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1132.588961] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.589753] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03323053-c6c2-4805-9bc1-1f0f8510e471 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.597791] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.598051] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec238dcc-6bfd-4930-910b-2a695ec4abe3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.604793] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1132.604793] env[63175]: value = "task-1248691" [ 1132.604793] env[63175]: _type = "Task" [ 1132.604793] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.613429] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.114248] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248691, 'name': PowerOffVM_Task, 'duration_secs': 0.185452} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.114471] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.114643] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1133.114888] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-788178dc-94aa-4083-8980-b1e52ba68381 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.177337] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.177565] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Deleting contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.177754] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore2] 5d676662-2185-4ef2-b9ab-b4b907992566 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.178029] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7ba1dd8-824b-4c8f-a779-115b72cf563a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.184544] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1133.184544] env[63175]: value = "task-1248693" [ 1133.184544] env[63175]: _type = "Task" [ 1133.184544] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.191685] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248693, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.694383] env[63175]: DEBUG oslo_vmware.api [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248693, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136827} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.694752] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.694802] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Deleted contents of the VM from datastore datastore2 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.694985] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.695191] env[63175]: INFO nova.compute.manager [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1133.695435] env[63175]: DEBUG oslo.service.loopingcall [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.695633] env[63175]: DEBUG nova.compute.manager [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1133.695726] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1133.931536] env[63175]: DEBUG nova.compute.manager [req-7671af1b-238e-451c-a86b-dbf8265682cc req-179300cc-a166-4ddf-82bf-c99e0b781aeb service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Received event network-vif-deleted-e0d257f6-bdfe-42e0-a4c2-5d8425081a51 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1133.931761] env[63175]: INFO nova.compute.manager [req-7671af1b-238e-451c-a86b-dbf8265682cc req-179300cc-a166-4ddf-82bf-c99e0b781aeb service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Neutron deleted interface e0d257f6-bdfe-42e0-a4c2-5d8425081a51; detaching it from the instance and deleting it from the info cache [ 1133.931933] env[63175]: DEBUG nova.network.neutron [req-7671af1b-238e-451c-a86b-dbf8265682cc req-179300cc-a166-4ddf-82bf-c99e0b781aeb service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.412192] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.434967] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3780297d-bc3e-4af9-b5f5-e0298fb1b876 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.444222] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da147ac0-228a-448f-b565-3c15d7b619eb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.467885] env[63175]: DEBUG nova.compute.manager [req-7671af1b-238e-451c-a86b-dbf8265682cc req-179300cc-a166-4ddf-82bf-c99e0b781aeb service nova] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Detach interface failed, port_id=e0d257f6-bdfe-42e0-a4c2-5d8425081a51, reason: Instance 5d676662-2185-4ef2-b9ab-b4b907992566 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1134.914742] env[63175]: INFO nova.compute.manager [-] [instance: 5d676662-2185-4ef2-b9ab-b4b907992566] Took 1.22 seconds to deallocate network for instance. [ 1135.421141] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.421458] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.421691] env[63175]: DEBUG nova.objects.instance [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 5d676662-2185-4ef2-b9ab-b4b907992566 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.969715] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed01814-3aa9-469a-98ce-1d4d23568218 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.977570] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2a23d5-0e05-45ef-b939-25d8ab086828 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.008290] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afa9497-d73a-41cf-8f5c-017e7755f5d9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.015898] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd7ac37-d987-494e-aab1-e13c5ae0810b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.028962] env[63175]: DEBUG nova.compute.provider_tree [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.532299] env[63175]: DEBUG nova.scheduler.client.report [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1137.037982] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.055044] env[63175]: INFO nova.scheduler.client.report [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 5d676662-2185-4ef2-b9ab-b4b907992566 [ 1137.564709] env[63175]: DEBUG oslo_concurrency.lockutils [None req-3851b748-2185-4eae-addd-933a737033ae tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d676662-2185-4ef2-b9ab-b4b907992566" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.483s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.039983] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.040284] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.542790] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1140.066524] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.066800] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.068285] env[63175]: INFO nova.compute.claims [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1141.111198] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1cba4-8748-437a-a868-721ecd23ffa2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.118950] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9fd18b-08dd-4df5-9891-c8d919489186 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.149440] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735a8792-f8cf-4f78-b57f-73e20db515bd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.156086] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753fff0f-976e-4273-875f-4aca6a2e258f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.168650] env[63175]: DEBUG nova.compute.provider_tree [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.672266] env[63175]: DEBUG nova.scheduler.client.report [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1142.177108] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.110s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.177701] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1142.682353] env[63175]: DEBUG nova.compute.utils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1142.684145] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1142.684339] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1142.736093] env[63175]: DEBUG nova.policy [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0afbf4d637af4d899945d238f8852552', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea5b7d92029c40fd824817c25e39ac1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1142.977542] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Successfully created port: c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1143.187457] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1144.196660] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1144.220710] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.220969] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.221148] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.221350] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.221570] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.221706] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.221918] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.222093] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.222269] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.222437] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.222612] env[63175]: DEBUG nova.virt.hardware [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.223587] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23379e1f-a0ac-4d62-96bc-40708d72c957 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.231792] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42300094-b49a-49f4-88fe-7977bbc9693c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.360302] env[63175]: DEBUG nova.compute.manager [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Received event network-vif-plugged-c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1144.360555] env[63175]: DEBUG oslo_concurrency.lockutils [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] Acquiring lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.360717] env[63175]: DEBUG oslo_concurrency.lockutils [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.360895] env[63175]: DEBUG oslo_concurrency.lockutils [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.361075] env[63175]: DEBUG nova.compute.manager [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] No waiting events found dispatching network-vif-plugged-c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1144.361248] env[63175]: WARNING nova.compute.manager [req-a5bcfe3f-f577-4208-a47c-6bb386be6991 req-5d67c035-085b-4995-ab0b-a63bade61dad service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Received unexpected event network-vif-plugged-c5896b03-901d-405c-87e7-538701ee2046 for instance with vm_state building and task_state spawning. [ 1144.413414] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Successfully updated port: c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1144.916371] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.916567] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.916662] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1145.448356] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1145.569166] env[63175]: DEBUG nova.network.neutron [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Updating instance_info_cache with network_info: [{"id": "c5896b03-901d-405c-87e7-538701ee2046", "address": "fa:16:3e:d1:a8:1b", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5896b03-90", "ovs_interfaceid": "c5896b03-901d-405c-87e7-538701ee2046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.071750] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.072087] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Instance network_info: |[{"id": "c5896b03-901d-405c-87e7-538701ee2046", "address": "fa:16:3e:d1:a8:1b", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5896b03-90", "ovs_interfaceid": "c5896b03-901d-405c-87e7-538701ee2046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1146.072583] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:a8:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da0e5087-d65b-416f-90fe-beaa9c534ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5896b03-901d-405c-87e7-538701ee2046', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.079841] env[63175]: DEBUG oslo.service.loopingcall [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.080057] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1146.080283] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b81e511-bcbc-4572-a994-3ead375d0dfd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.100686] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1146.100686] env[63175]: value = "task-1248694" [ 1146.100686] env[63175]: _type = "Task" [ 1146.100686] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.107998] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248694, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.386194] env[63175]: DEBUG nova.compute.manager [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Received event network-changed-c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1146.386473] env[63175]: DEBUG nova.compute.manager [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Refreshing instance network info cache due to event network-changed-c5896b03-901d-405c-87e7-538701ee2046. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1146.386671] env[63175]: DEBUG oslo_concurrency.lockutils [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] Acquiring lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.386827] env[63175]: DEBUG oslo_concurrency.lockutils [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] Acquired lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.386995] env[63175]: DEBUG nova.network.neutron [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Refreshing network info cache for port c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1146.610921] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248694, 'name': CreateVM_Task, 'duration_secs': 0.298074} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.611229] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1146.611805] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.611973] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.612336] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1146.612560] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-319562d2-6968-448d-9b61-b58b6c6ad357 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.616715] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1146.616715] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525953df-6348-d9bf-8029-2e967520be43" [ 1146.616715] env[63175]: _type = "Task" [ 1146.616715] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.623957] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525953df-6348-d9bf-8029-2e967520be43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.071583] env[63175]: DEBUG nova.network.neutron [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Updated VIF entry in instance network info cache for port c5896b03-901d-405c-87e7-538701ee2046. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1147.071955] env[63175]: DEBUG nova.network.neutron [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Updating instance_info_cache with network_info: [{"id": "c5896b03-901d-405c-87e7-538701ee2046", "address": "fa:16:3e:d1:a8:1b", "network": {"id": "95a6b873-c45d-4e57-ab83-dd35eb6cac38", "bridge": "br-int", "label": "tempest-ServersTestJSON-1853116024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea5b7d92029c40fd824817c25e39ac1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da0e5087-d65b-416f-90fe-beaa9c534ad3", "external-id": "nsx-vlan-transportzone-522", "segmentation_id": 522, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5896b03-90", "ovs_interfaceid": "c5896b03-901d-405c-87e7-538701ee2046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.127555] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]525953df-6348-d9bf-8029-2e967520be43, 'name': SearchDatastore_Task, 'duration_secs': 0.013757} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.127966] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.128301] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1147.128622] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.128860] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.129150] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1147.129483] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df209f91-d1bf-49d0-9c80-610f275af060 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.137682] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1147.137944] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1147.138797] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4515f2b3-e315-4092-b5da-206bce5d95bc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.144163] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1147.144163] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524d877e-1492-2a72-cad1-098c22bce575" [ 1147.144163] env[63175]: _type = "Task" [ 1147.144163] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.152219] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524d877e-1492-2a72-cad1-098c22bce575, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.574753] env[63175]: DEBUG oslo_concurrency.lockutils [req-10150b15-d411-4fab-b54a-3aa6629742d9 req-c6361afd-18c9-43fe-9a72-275e67d342c3 service nova] Releasing lock "refresh_cache-c8b2ba6d-f82e-48dc-8955-91597e7a1be9" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.654976] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]524d877e-1492-2a72-cad1-098c22bce575, 'name': SearchDatastore_Task, 'duration_secs': 0.008315} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.655687] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d8b356a-a7a4-4f12-a29f-b42257d44de4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.660537] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1147.660537] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5227e150-c405-81eb-ac52-10acfe96e7d9" [ 1147.660537] env[63175]: _type = "Task" [ 1147.660537] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.668617] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5227e150-c405-81eb-ac52-10acfe96e7d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.170789] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5227e150-c405-81eb-ac52-10acfe96e7d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009089} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.171082] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.171351] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] c8b2ba6d-f82e-48dc-8955-91597e7a1be9/c8b2ba6d-f82e-48dc-8955-91597e7a1be9.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1148.171610] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6c90f2b-fcb1-43ff-bcae-2376c685a812 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.178093] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1148.178093] env[63175]: value = "task-1248695" [ 1148.178093] env[63175]: _type = "Task" [ 1148.178093] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.185965] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.687878] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459445} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.688261] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] c8b2ba6d-f82e-48dc-8955-91597e7a1be9/c8b2ba6d-f82e-48dc-8955-91597e7a1be9.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1148.688402] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1148.688622] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52e10409-0dcc-4f97-be0b-445343ef4def {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.695901] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1148.695901] env[63175]: value = "task-1248696" [ 1148.695901] env[63175]: _type = "Task" [ 1148.695901] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.703143] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248696, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.205361] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248696, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061703} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.205622] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1149.206413] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24059857-3060-4106-a4fa-4d63bfa03fee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.226942] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c8b2ba6d-f82e-48dc-8955-91597e7a1be9/c8b2ba6d-f82e-48dc-8955-91597e7a1be9.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.227185] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09e2f11f-9542-4f51-b7fb-e49477b8b11c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.246559] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1149.246559] env[63175]: value = "task-1248697" [ 1149.246559] env[63175]: _type = "Task" [ 1149.246559] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.253696] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248697, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.756502] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248697, 'name': ReconfigVM_Task, 'duration_secs': 0.33592} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.756801] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c8b2ba6d-f82e-48dc-8955-91597e7a1be9/c8b2ba6d-f82e-48dc-8955-91597e7a1be9.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.757478] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8943e38e-df6e-4472-a7a2-e5e34d22c8a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.763422] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1149.763422] env[63175]: value = "task-1248698" [ 1149.763422] env[63175]: _type = "Task" [ 1149.763422] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.770682] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248698, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.272773] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248698, 'name': Rename_Task, 'duration_secs': 0.130245} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.273099] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1150.273349] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1925221c-0aa3-45a5-9831-07831cbc12f1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.279355] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1150.279355] env[63175]: value = "task-1248699" [ 1150.279355] env[63175]: _type = "Task" [ 1150.279355] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.288884] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.789792] env[63175]: DEBUG oslo_vmware.api [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248699, 'name': PowerOnVM_Task, 'duration_secs': 0.430225} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.790184] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1150.790333] env[63175]: INFO nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Took 6.59 seconds to spawn the instance on the hypervisor. [ 1150.790521] env[63175]: DEBUG nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1150.791312] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbfe94e-173c-4444-b347-5540c5091a2c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.308905] env[63175]: INFO nova.compute.manager [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Took 11.26 seconds to build instance. [ 1151.810939] env[63175]: DEBUG oslo_concurrency.lockutils [None req-6064a59f-31b2-41fc-83bb-aa2e07fcdbb4 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.770s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.434058] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.434284] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.434411] env[63175]: DEBUG nova.compute.manager [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1152.435673] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0123dae-9e85-4a17-b093-f04a53e30f08 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.442092] env[63175]: DEBUG nova.compute.manager [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63175) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1152.442633] env[63175]: DEBUG nova.objects.instance [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'flavor' on Instance uuid c8b2ba6d-f82e-48dc-8955-91597e7a1be9 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.449705] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1153.450129] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-616c7dd4-d637-4229-88c4-5906b8287cce {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.457582] env[63175]: DEBUG oslo_vmware.api [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1153.457582] env[63175]: value = "task-1248700" [ 1153.457582] env[63175]: _type = "Task" [ 1153.457582] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.465200] env[63175]: DEBUG oslo_vmware.api [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.968067] env[63175]: DEBUG oslo_vmware.api [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248700, 'name': PowerOffVM_Task, 'duration_secs': 0.197144} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.968067] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1153.968295] env[63175]: DEBUG nova.compute.manager [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1153.969071] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862c8a88-aec8-4550-a23f-c82257168844 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.480108] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1c5a24f8-fa5f-4982-889a-64e89c959ced tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.046s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.045839] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.046213] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.046507] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.046699] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.046876] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.049040] env[63175]: INFO nova.compute.manager [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Terminating instance [ 1155.553246] env[63175]: DEBUG nova.compute.manager [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1155.553678] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1155.554428] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d238aff-b4a8-4269-a26b-28342ae6a909 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.562379] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1155.562614] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8e4ac60-39df-44c5-bc52-91e96a95e68a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.627767] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1155.627976] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1155.628180] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore1] c8b2ba6d-f82e-48dc-8955-91597e7a1be9 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1155.628466] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d4e3e02-42dd-4a26-b1c2-55ab5ea85c90 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.634147] env[63175]: DEBUG oslo_vmware.api [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1155.634147] env[63175]: value = "task-1248702" [ 1155.634147] env[63175]: _type = "Task" [ 1155.634147] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.642894] env[63175]: DEBUG oslo_vmware.api [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.143036] env[63175]: DEBUG oslo_vmware.api [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123323} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.143306] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.143501] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1156.143682] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1156.143862] env[63175]: INFO nova.compute.manager [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1156.144119] env[63175]: DEBUG oslo.service.loopingcall [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.144338] env[63175]: DEBUG nova.compute.manager [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1156.144438] env[63175]: DEBUG nova.network.neutron [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1156.368173] env[63175]: DEBUG nova.compute.manager [req-83418a55-06da-45d2-8bfd-31128f92824e req-5ac39c00-1b02-437a-bf80-ec29bf553042 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Received event network-vif-deleted-c5896b03-901d-405c-87e7-538701ee2046 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1156.368378] env[63175]: INFO nova.compute.manager [req-83418a55-06da-45d2-8bfd-31128f92824e req-5ac39c00-1b02-437a-bf80-ec29bf553042 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Neutron deleted interface c5896b03-901d-405c-87e7-538701ee2046; detaching it from the instance and deleting it from the info cache [ 1156.368559] env[63175]: DEBUG nova.network.neutron [req-83418a55-06da-45d2-8bfd-31128f92824e req-5ac39c00-1b02-437a-bf80-ec29bf553042 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.844856] env[63175]: DEBUG nova.network.neutron [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.871940] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5fd49df-6fe1-4370-b003-c9c0db98b23f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.884592] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986fb7dc-a9d9-427e-abf1-b5555623034f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.909200] env[63175]: DEBUG nova.compute.manager [req-83418a55-06da-45d2-8bfd-31128f92824e req-5ac39c00-1b02-437a-bf80-ec29bf553042 service nova] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Detach interface failed, port_id=c5896b03-901d-405c-87e7-538701ee2046, reason: Instance c8b2ba6d-f82e-48dc-8955-91597e7a1be9 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1157.348265] env[63175]: INFO nova.compute.manager [-] [instance: c8b2ba6d-f82e-48dc-8955-91597e7a1be9] Took 1.20 seconds to deallocate network for instance. [ 1157.855320] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.855722] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.856082] env[63175]: DEBUG nova.objects.instance [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid c8b2ba6d-f82e-48dc-8955-91597e7a1be9 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.375542] env[63175]: DEBUG nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Refreshing inventories for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1158.388957] env[63175]: DEBUG nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Updating ProviderTree inventory for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1158.389222] env[63175]: DEBUG nova.compute.provider_tree [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Updating inventory in ProviderTree for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1158.400435] env[63175]: DEBUG nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Refreshing aggregate associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, aggregates: None {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1158.416292] env[63175]: DEBUG nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Refreshing trait associations for resource provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=63175) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1158.451254] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de82d2e6-56ba-4efc-926b-a2d733d3659d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.458905] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb1d215-2b35-4468-be8b-5bbd569f568c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.489012] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bb180a-c408-4f3b-8325-e09c98692720 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.495747] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d28409-c102-4f2b-8eb8-582b14f53c0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.508374] env[63175]: DEBUG nova.compute.provider_tree [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.012040] env[63175]: DEBUG nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1159.517137] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.539429] env[63175]: INFO nova.scheduler.client.report [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance c8b2ba6d-f82e-48dc-8955-91597e7a1be9 [ 1160.048259] env[63175]: DEBUG oslo_concurrency.lockutils [None req-59742675-e41a-4a27-9a4d-d5aee55317a5 tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "c8b2ba6d-f82e-48dc-8955-91597e7a1be9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.002s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.817223] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.817490] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.817723] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.817922] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.818120] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.820437] env[63175]: INFO nova.compute.manager [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Terminating instance [ 1161.324344] env[63175]: DEBUG nova.compute.manager [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1161.324724] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1161.325557] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15533024-2b25-4e90-8b20-124aadf401ac {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.333505] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.333739] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17c7940f-e4fb-4b6a-8d2a-8ce09046589b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.339150] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1161.339150] env[63175]: value = "task-1248703" [ 1161.339150] env[63175]: _type = "Task" [ 1161.339150] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.346709] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.848659] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248703, 'name': PowerOffVM_Task, 'duration_secs': 0.172773} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.848874] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1161.849065] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1161.849317] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a734677-7363-4d75-9cfb-dbcb791eb1a1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.915728] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1161.915956] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1161.916168] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleting the datastore file [datastore1] 5d4d277e-ec2d-40ed-b16a-6d87965888e2 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.916469] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57cbf46d-b27a-4a70-85ce-739f7bca10d5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.924280] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for the task: (returnval){ [ 1161.924280] env[63175]: value = "task-1248705" [ 1161.924280] env[63175]: _type = "Task" [ 1161.924280] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.931466] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.433975] env[63175]: DEBUG oslo_vmware.api [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Task: {'id': task-1248705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122857} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.434394] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.434474] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1162.434652] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.434836] env[63175]: INFO nova.compute.manager [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1162.435099] env[63175]: DEBUG oslo.service.loopingcall [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1162.435300] env[63175]: DEBUG nova.compute.manager [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1162.435395] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1162.660732] env[63175]: DEBUG nova.compute.manager [req-89f407fb-4cd2-4eb1-9b30-07d7a8e557f3 req-3bfe5304-2024-4dc9-b1b4-10c2fc4904b1 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Received event network-vif-deleted-72df71c2-faab-4cc4-b2c3-e730bb625453 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1162.660960] env[63175]: INFO nova.compute.manager [req-89f407fb-4cd2-4eb1-9b30-07d7a8e557f3 req-3bfe5304-2024-4dc9-b1b4-10c2fc4904b1 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Neutron deleted interface 72df71c2-faab-4cc4-b2c3-e730bb625453; detaching it from the instance and deleting it from the info cache [ 1162.661189] env[63175]: DEBUG nova.network.neutron [req-89f407fb-4cd2-4eb1-9b30-07d7a8e557f3 req-3bfe5304-2024-4dc9-b1b4-10c2fc4904b1 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.139848] env[63175]: DEBUG nova.network.neutron [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.163119] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e5216c5-56d9-4ea7-8baa-c8c4c37e8830 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.173403] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666ff779-668b-4cd0-81f0-4ecb700f217a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.195290] env[63175]: DEBUG nova.compute.manager [req-89f407fb-4cd2-4eb1-9b30-07d7a8e557f3 req-3bfe5304-2024-4dc9-b1b4-10c2fc4904b1 service nova] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Detach interface failed, port_id=72df71c2-faab-4cc4-b2c3-e730bb625453, reason: Instance 5d4d277e-ec2d-40ed-b16a-6d87965888e2 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1163.642667] env[63175]: INFO nova.compute.manager [-] [instance: 5d4d277e-ec2d-40ed-b16a-6d87965888e2] Took 1.21 seconds to deallocate network for instance. [ 1164.149418] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.149638] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.149866] env[63175]: DEBUG nova.objects.instance [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lazy-loading 'resources' on Instance uuid 5d4d277e-ec2d-40ed-b16a-6d87965888e2 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.681813] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e4cf68-b01e-427c-b28b-a2e8de19d698 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.689315] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963cca18-4f51-4b6f-aa2c-13d45b1fdc9c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.718922] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4c2353-fa7d-416c-84f6-c7c98884cbb1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.725758] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a780bb-b89b-4afb-aea8-e44d41ef86fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.738454] env[63175]: DEBUG nova.compute.provider_tree [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.241239] env[63175]: DEBUG nova.scheduler.client.report [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1165.745896] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.596s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.764286] env[63175]: INFO nova.scheduler.client.report [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Deleted allocations for instance 5d4d277e-ec2d-40ed-b16a-6d87965888e2 [ 1166.271379] env[63175]: DEBUG oslo_concurrency.lockutils [None req-d0d2c3fc-39bc-4d88-974a-c162358e1b7e tempest-ServersTestJSON-201069960 tempest-ServersTestJSON-201069960-project-member] Lock "5d4d277e-ec2d-40ed-b16a-6d87965888e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.454s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.398217] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.398564] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1167.398564] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1168.404909] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1168.405169] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.405316] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.405459] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.405645] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.405793] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.405942] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.406077] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1168.406226] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1168.909478] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.909718] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.909890] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.910061] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1168.910947] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436f2172-476c-4471-af64-fcf33cc92ca4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.919219] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b731fdae-0a6e-4497-9dfd-8f209568fd07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.933604] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70bec63-ff68-4611-9d6c-468884f18bc6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.940143] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce89448a-f0eb-4cd2-8a19-24b8c68e3d7f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.971146] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181229MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1168.971322] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.971523] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.991348] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1169.991567] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1170.004474] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f30f726-f37a-4cdc-8cce-20ee98114da4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.013121] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcccc956-0538-46d3-aa25-bc1d732af77b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.041784] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76947b42-e24d-4584-b0ae-45681c0cab36 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.048662] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639dc691-a19a-4004-8269-b1c6d8a6092a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.062527] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.565237] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1171.070851] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1171.071357] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.100s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.462280] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.462527] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.964939] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1172.486723] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.487025] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.488548] env[63175]: INFO nova.compute.claims [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1173.521210] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbaae1d1-6e48-4df3-ae68-729596ea3551 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.528867] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0a3ba0-2e4b-4d75-9ba3-4c1baee44e8e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.558616] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1740fca9-d4e3-4f7a-8f84-a7b77f27abe0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.565353] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2adf9df-48ec-43ad-aecd-bdf4de40c779 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.578282] env[63175]: DEBUG nova.compute.provider_tree [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.081662] env[63175]: DEBUG nova.scheduler.client.report [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1174.586260] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.099s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.586833] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1175.092106] env[63175]: DEBUG nova.compute.utils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.093594] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1175.093712] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1175.139071] env[63175]: DEBUG nova.policy [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9a28543af5f433285d117a0b0d5fb44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca6ed794d781428982398dcfb3998f08', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1175.367658] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Successfully created port: 21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1175.597303] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1176.607365] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1176.631621] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1176.631877] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1176.632054] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1176.632291] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1176.632459] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1176.632613] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1176.632827] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1176.632991] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1176.633187] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1176.633357] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1176.633531] env[63175]: DEBUG nova.virt.hardware [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1176.634474] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6562e0-a4ae-4d65-a9cf-2541d1772ee6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.642576] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08a3762-dcf8-4024-912d-6ca4b6096efd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.725407] env[63175]: DEBUG nova.compute.manager [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1176.725644] env[63175]: DEBUG oslo_concurrency.lockutils [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.725851] env[63175]: DEBUG oslo_concurrency.lockutils [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.726046] env[63175]: DEBUG oslo_concurrency.lockutils [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.726238] env[63175]: DEBUG nova.compute.manager [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] No waiting events found dispatching network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1176.726410] env[63175]: WARNING nova.compute.manager [req-54069c8e-5c85-471a-ae7e-3a4564325fbb req-fbcf3d8a-054b-4078-8e59-932c6f7dc68c service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received unexpected event network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e for instance with vm_state building and task_state spawning. [ 1176.805780] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Successfully updated port: 21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1177.309055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.309221] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.309374] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1177.843399] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1177.963991] env[63175]: DEBUG nova.network.neutron [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.466932] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.467278] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance network_info: |[{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1178.467713] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:06:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21577a64-90af-48ef-b945-bb0bed6cc80e', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1178.474970] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating folder: Project (ca6ed794d781428982398dcfb3998f08). Parent ref: group-v268956. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1178.475250] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99db2868-d83f-447a-9e6d-5c3b43f5d5b0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.488041] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created folder: Project (ca6ed794d781428982398dcfb3998f08) in parent group-v268956. [ 1178.488226] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating folder: Instances. Parent ref: group-v269118. {{(pid=63175) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1178.488761] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da711e37-8392-4166-8d50-29d8e8be4350 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.497914] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created folder: Instances in parent group-v269118. [ 1178.498157] env[63175]: DEBUG oslo.service.loopingcall [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.498340] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1178.498534] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc7756a8-ae0e-44b0-b18f-90a3a76d679d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.516563] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1178.516563] env[63175]: value = "task-1248708" [ 1178.516563] env[63175]: _type = "Task" [ 1178.516563] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.523558] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248708, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.752209] env[63175]: DEBUG nova.compute.manager [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1178.752521] env[63175]: DEBUG nova.compute.manager [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing instance network info cache due to event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1178.752769] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.752922] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.753106] env[63175]: DEBUG nova.network.neutron [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1179.026128] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248708, 'name': CreateVM_Task, 'duration_secs': 0.334938} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.026447] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1179.026949] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.027130] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.027451] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1179.027720] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4f3fcc0-22ca-4b6c-a570-2c3c385234e7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.031733] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1179.031733] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8a860-70da-6b0d-9cce-84a4daa8be29" [ 1179.031733] env[63175]: _type = "Task" [ 1179.031733] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.038839] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8a860-70da-6b0d-9cce-84a4daa8be29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.432418] env[63175]: DEBUG nova.network.neutron [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updated VIF entry in instance network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1179.432764] env[63175]: DEBUG nova.network.neutron [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.542300] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52a8a860-70da-6b0d-9cce-84a4daa8be29, 'name': SearchDatastore_Task, 'duration_secs': 0.011721} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.542568] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.542807] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1179.543051] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.543209] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.543394] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1179.543652] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c255dcc-38d7-4aa7-a9cc-61588d3d9da4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.551491] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1179.551679] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1179.552368] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9381e9ea-0792-4615-a4c8-5502b875ee91 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.557097] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1179.557097] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526dd058-f223-03eb-01dc-f3a741fa72ca" [ 1179.557097] env[63175]: _type = "Task" [ 1179.557097] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.564918] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526dd058-f223-03eb-01dc-f3a741fa72ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.935957] env[63175]: DEBUG oslo_concurrency.lockutils [req-6c39be5e-bbcb-4bc8-85ad-af3196dab496 req-0a27f3bb-721d-4249-ad79-eaefa4afbcb6 service nova] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.067457] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]526dd058-f223-03eb-01dc-f3a741fa72ca, 'name': SearchDatastore_Task, 'duration_secs': 0.00863} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.068243] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc87f130-aefe-4bd3-83dd-d8b00cb15350 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.073459] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1180.073459] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523f7e3b-aa89-ea50-23cd-5ba2decbdc04" [ 1180.073459] env[63175]: _type = "Task" [ 1180.073459] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.081790] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523f7e3b-aa89-ea50-23cd-5ba2decbdc04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.585051] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]523f7e3b-aa89-ea50-23cd-5ba2decbdc04, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.585238] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.585514] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1180.585772] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1954ede-5c13-4761-9604-3fa60197e5b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.592557] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1180.592557] env[63175]: value = "task-1248709" [ 1180.592557] env[63175]: _type = "Task" [ 1180.592557] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.599882] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.102959] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440435} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.103345] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1181.103553] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1181.103848] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f315074-4c44-481f-a3cb-8f2e86ab54cb {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.110429] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1181.110429] env[63175]: value = "task-1248710" [ 1181.110429] env[63175]: _type = "Task" [ 1181.110429] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.117820] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.620013] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065415} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.620326] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1181.621084] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a58a00c-6130-4830-9d10-e880f0729db1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.642299] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.642520] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31cc3ea6-1bcb-4c9c-9e79-7b0d71f37d25 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.662022] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1181.662022] env[63175]: value = "task-1248711" [ 1181.662022] env[63175]: _type = "Task" [ 1181.662022] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.669824] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.171645] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248711, 'name': ReconfigVM_Task, 'duration_secs': 0.277971} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.172035] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Reconfigured VM instance instance-0000006b to attach disk [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.172563] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2c062db-4ef9-468e-8b72-c4f7f1b571ca {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.178783] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1182.178783] env[63175]: value = "task-1248712" [ 1182.178783] env[63175]: _type = "Task" [ 1182.178783] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.190165] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248712, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.688478] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248712, 'name': Rename_Task, 'duration_secs': 0.134435} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.688739] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.689016] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70881c01-34af-4def-a657-070135530b23 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.694728] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1182.694728] env[63175]: value = "task-1248713" [ 1182.694728] env[63175]: _type = "Task" [ 1182.694728] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.702393] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.204398] env[63175]: DEBUG oslo_vmware.api [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248713, 'name': PowerOnVM_Task, 'duration_secs': 0.46273} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.204806] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1183.204861] env[63175]: INFO nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Took 6.60 seconds to spawn the instance on the hypervisor. [ 1183.205052] env[63175]: DEBUG nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1183.205802] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b3771f-0f9f-454f-9de5-620b3e7ea459 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.722857] env[63175]: INFO nova.compute.manager [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Took 11.25 seconds to build instance. [ 1184.225318] env[63175]: DEBUG oslo_concurrency.lockutils [None req-24cf1f40-3c0f-421c-951c-5fcd3aa1c5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.763s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.944736] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "92a9e153-06fb-420e-a96d-874ace510c5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.944999] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.447369] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Starting instance... {{(pid=63175) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1185.967674] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.967962] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.969438] env[63175]: INFO nova.compute.claims [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1187.014726] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d338e23-bd98-4ac6-9584-e0e54c831291 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.022099] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7814e15-f2db-434b-b5c0-03fc432f38da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.050350] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8928a70-7ed1-47ac-bf4b-8e9671fd10e2 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.056766] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a24ca0-695d-441f-b133-137636c84eaa {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.070026] env[63175]: DEBUG nova.compute.provider_tree [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.572662] env[63175]: DEBUG nova.scheduler.client.report [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1188.077474] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.078092] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Start building networks asynchronously for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1188.583240] env[63175]: DEBUG nova.compute.utils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Using /dev/sd instead of None {{(pid=63175) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1188.584628] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Allocating IP information in the background. {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1188.584810] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] allocate_for_instance() {{(pid=63175) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1188.634550] env[63175]: DEBUG nova.policy [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9a28543af5f433285d117a0b0d5fb44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca6ed794d781428982398dcfb3998f08', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63175) authorize /opt/stack/nova/nova/policy.py:201}} [ 1188.879471] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Successfully created port: 03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1189.087914] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Start building block device mappings for instance. {{(pid=63175) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1190.098768] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Start spawning the instance on the hypervisor. {{(pid=63175) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1190.123411] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T14:42:48Z,direct_url=,disk_format='vmdk',id=8e9c4b2c-a2a7-4856-ba6d-ed187f088508,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='e0fc85e1ac6e4bf088fba7e5b174c44f',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T14:42:49Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1190.123713] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1190.123880] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1190.124086] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1190.124238] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1190.124390] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1190.124604] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1190.124779] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1190.124945] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1190.125202] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1190.125393] env[63175]: DEBUG nova.virt.hardware [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1190.126265] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973b80d3-472e-40b9-ad0a-862990a0d2d4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.134037] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea21602-39e4-4a71-af65-27ce6e8da717 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.242284] env[63175]: DEBUG nova.compute.manager [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Received event network-vif-plugged-03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1190.242284] env[63175]: DEBUG oslo_concurrency.lockutils [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] Acquiring lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.242284] env[63175]: DEBUG oslo_concurrency.lockutils [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] Lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.242284] env[63175]: DEBUG oslo_concurrency.lockutils [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] Lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.242284] env[63175]: DEBUG nova.compute.manager [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] No waiting events found dispatching network-vif-plugged-03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1190.242284] env[63175]: WARNING nova.compute.manager [req-f7ea6eb4-3c1d-47f7-9c97-8dbb712b4e73 req-79fb2d1b-65cd-44f9-8382-69bbd7497b46 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Received unexpected event network-vif-plugged-03f5abc1-a5ae-43d6-8523-18e2f4996ee8 for instance with vm_state building and task_state spawning. [ 1190.319965] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Successfully updated port: 03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1190.822281] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.822437] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.822592] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1191.352015] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Instance cache missing network info. {{(pid=63175) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1191.467435] env[63175]: DEBUG nova.network.neutron [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Updating instance_info_cache with network_info: [{"id": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "address": "fa:16:3e:52:b3:ca", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f5abc1-a5", "ovs_interfaceid": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.970179] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.970513] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Instance network_info: |[{"id": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "address": "fa:16:3e:52:b3:ca", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f5abc1-a5", "ovs_interfaceid": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63175) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1191.970950] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:b3:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03f5abc1-a5ae-43d6-8523-18e2f4996ee8', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.978303] env[63175]: DEBUG oslo.service.loopingcall [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.978509] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.979153] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-627feacd-e145-43b0-8f90-5fc8300be370 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.998833] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.998833] env[63175]: value = "task-1248714" [ 1191.998833] env[63175]: _type = "Task" [ 1191.998833] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.005949] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248714, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.267875] env[63175]: DEBUG nova.compute.manager [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Received event network-changed-03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1192.268179] env[63175]: DEBUG nova.compute.manager [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Refreshing instance network info cache due to event network-changed-03f5abc1-a5ae-43d6-8523-18e2f4996ee8. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1192.268493] env[63175]: DEBUG oslo_concurrency.lockutils [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] Acquiring lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.268721] env[63175]: DEBUG oslo_concurrency.lockutils [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] Acquired lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.268912] env[63175]: DEBUG nova.network.neutron [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Refreshing network info cache for port 03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1192.508936] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248714, 'name': CreateVM_Task, 'duration_secs': 0.289419} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.509347] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.509742] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.509910] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.510246] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1192.510494] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3b86f39-7b69-413c-9b90-3b3d30e7e345 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.514743] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1192.514743] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52546f96-3b07-c28e-8b48-ea67e77a0680" [ 1192.514743] env[63175]: _type = "Task" [ 1192.514743] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.521983] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52546f96-3b07-c28e-8b48-ea67e77a0680, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.942920] env[63175]: DEBUG nova.network.neutron [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Updated VIF entry in instance network info cache for port 03f5abc1-a5ae-43d6-8523-18e2f4996ee8. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1192.943324] env[63175]: DEBUG nova.network.neutron [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Updating instance_info_cache with network_info: [{"id": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "address": "fa:16:3e:52:b3:ca", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f5abc1-a5", "ovs_interfaceid": "03f5abc1-a5ae-43d6-8523-18e2f4996ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.025053] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52546f96-3b07-c28e-8b48-ea67e77a0680, 'name': SearchDatastore_Task, 'duration_secs': 0.009682} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.025198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.025372] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Processing image 8e9c4b2c-a2a7-4856-ba6d-ed187f088508 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1193.025607] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.025760] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.025943] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1193.026392] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d70027ac-202a-43f8-acf3-c13064518fbf {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.033832] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1193.034017] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1193.034734] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb622ac5-bfb4-4a1e-bcae-a9e915772150 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.039348] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1193.039348] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5272b545-51c0-0c32-abec-fde417ef307b" [ 1193.039348] env[63175]: _type = "Task" [ 1193.039348] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.046711] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5272b545-51c0-0c32-abec-fde417ef307b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.446245] env[63175]: DEBUG oslo_concurrency.lockutils [req-fd33a932-8d76-4b8c-9c72-b574e4253583 req-3cff75e3-db2b-45a3-8c41-02ffa2c42c29 service nova] Releasing lock "refresh_cache-92a9e153-06fb-420e-a96d-874ace510c5b" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.549425] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5272b545-51c0-0c32-abec-fde417ef307b, 'name': SearchDatastore_Task, 'duration_secs': 0.007704} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.550189] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6bd3a0b-86df-47f1-a19a-3a03fce9bac9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.555343] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1193.555343] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cf2784-7551-525d-bafe-65a400067c34" [ 1193.555343] env[63175]: _type = "Task" [ 1193.555343] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.562444] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cf2784-7551-525d-bafe-65a400067c34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.065068] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52cf2784-7551-525d-bafe-65a400067c34, 'name': SearchDatastore_Task, 'duration_secs': 0.008562} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.065294] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.065555] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 92a9e153-06fb-420e-a96d-874ace510c5b/92a9e153-06fb-420e-a96d-874ace510c5b.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1194.065799] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6fd58eb9-f62e-47a4-87f5-eefc60d22699 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.072158] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1194.072158] env[63175]: value = "task-1248715" [ 1194.072158] env[63175]: _type = "Task" [ 1194.072158] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.080275] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.581159] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.409858} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.581599] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8e9c4b2c-a2a7-4856-ba6d-ed187f088508/8e9c4b2c-a2a7-4856-ba6d-ed187f088508.vmdk to [datastore1] 92a9e153-06fb-420e-a96d-874ace510c5b/92a9e153-06fb-420e-a96d-874ace510c5b.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1194.581675] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Extending root virtual disk to 1048576 {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.581858] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49c43d83-4d9f-44d2-8975-2672d597b169 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.588080] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1194.588080] env[63175]: value = "task-1248716" [ 1194.588080] env[63175]: _type = "Task" [ 1194.588080] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.594872] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248716, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.100125] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248716, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068163} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.100488] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Extended root virtual disk {{(pid=63175) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1195.101535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bc0de7-2f8e-4d12-9762-a1324ad92df4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.124465] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 92a9e153-06fb-420e-a96d-874ace510c5b/92a9e153-06fb-420e-a96d-874ace510c5b.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1195.124704] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f6e1ea3-425b-4ee7-a49b-4292a21260b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.143240] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1195.143240] env[63175]: value = "task-1248717" [ 1195.143240] env[63175]: _type = "Task" [ 1195.143240] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.150489] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248717, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.652507] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248717, 'name': ReconfigVM_Task, 'duration_secs': 0.312272} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.654045] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 92a9e153-06fb-420e-a96d-874ace510c5b/92a9e153-06fb-420e-a96d-874ace510c5b.vmdk or device None with type sparse {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.654045] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6224ee65-8c41-41d0-a3cd-ea6edfc9c47c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.660029] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1195.660029] env[63175]: value = "task-1248718" [ 1195.660029] env[63175]: _type = "Task" [ 1195.660029] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.667421] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248718, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.170334] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248718, 'name': Rename_Task, 'duration_secs': 0.160513} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.170632] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1196.170895] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4996cbab-c071-425a-bc44-c3a3b060cde8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.178055] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1196.178055] env[63175]: value = "task-1248719" [ 1196.178055] env[63175]: _type = "Task" [ 1196.178055] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.185340] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.687278] env[63175]: DEBUG oslo_vmware.api [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248719, 'name': PowerOnVM_Task, 'duration_secs': 0.439125} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.687665] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1196.687789] env[63175]: INFO nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Took 6.59 seconds to spawn the instance on the hypervisor. [ 1196.687976] env[63175]: DEBUG nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1196.688738] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e20f77b-1897-4f7b-b546-33ef08ef9a33 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.203754] env[63175]: INFO nova.compute.manager [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Took 11.25 seconds to build instance. [ 1197.705225] env[63175]: DEBUG oslo_concurrency.lockutils [None req-539a4ba1-9d51-45e7-93a8-df24868b4c13 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.760s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.099109] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "92a9e153-06fb-420e-a96d-874ace510c5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.099109] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.099340] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.099523] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.099697] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.103263] env[63175]: INFO nova.compute.manager [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Terminating instance [ 1198.607061] env[63175]: DEBUG nova.compute.manager [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1198.607311] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1198.608587] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c712d1-db20-4e8b-8f1d-72e44711c4de {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.616248] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1198.616486] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6558fb06-c4c8-4acd-a788-c33183465c0b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.622909] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1198.622909] env[63175]: value = "task-1248720" [ 1198.622909] env[63175]: _type = "Task" [ 1198.622909] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.630537] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.132229] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248720, 'name': PowerOffVM_Task, 'duration_secs': 0.185537} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.132606] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1199.132654] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1199.132904] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74d5e516-a9a8-47bb-8fa5-c72ff8af52b6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.194521] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1199.194747] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1199.194936] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleting the datastore file [datastore1] 92a9e153-06fb-420e-a96d-874ace510c5b {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1199.195222] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8f52eda-bbca-458b-87a4-fbcdf139ff96 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.201397] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1199.201397] env[63175]: value = "task-1248722" [ 1199.201397] env[63175]: _type = "Task" [ 1199.201397] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.209020] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.710189] env[63175]: DEBUG oslo_vmware.api [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134791} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.710462] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1199.710660] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1199.710824] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1199.711013] env[63175]: INFO nova.compute.manager [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1199.711268] env[63175]: DEBUG oslo.service.loopingcall [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1199.711503] env[63175]: DEBUG nova.compute.manager [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1199.711628] env[63175]: DEBUG nova.network.neutron [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1199.947469] env[63175]: DEBUG nova.compute.manager [req-7034900d-4d5f-438f-b6ce-d9604503b9e3 req-f3012049-94fa-4195-9458-eb10aaff5e98 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Received event network-vif-deleted-03f5abc1-a5ae-43d6-8523-18e2f4996ee8 {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1199.947677] env[63175]: INFO nova.compute.manager [req-7034900d-4d5f-438f-b6ce-d9604503b9e3 req-f3012049-94fa-4195-9458-eb10aaff5e98 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Neutron deleted interface 03f5abc1-a5ae-43d6-8523-18e2f4996ee8; detaching it from the instance and deleting it from the info cache [ 1199.947899] env[63175]: DEBUG nova.network.neutron [req-7034900d-4d5f-438f-b6ce-d9604503b9e3 req-f3012049-94fa-4195-9458-eb10aaff5e98 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.426831] env[63175]: DEBUG nova.network.neutron [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.450257] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a14b9d94-ec04-4c99-8165-ab6aa3be5ab0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.460593] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c632306f-49b0-4e31-842a-47a928434331 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.483413] env[63175]: DEBUG nova.compute.manager [req-7034900d-4d5f-438f-b6ce-d9604503b9e3 req-f3012049-94fa-4195-9458-eb10aaff5e98 service nova] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Detach interface failed, port_id=03f5abc1-a5ae-43d6-8523-18e2f4996ee8, reason: Instance 92a9e153-06fb-420e-a96d-874ace510c5b could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1200.929195] env[63175]: INFO nova.compute.manager [-] [instance: 92a9e153-06fb-420e-a96d-874ace510c5b] Took 1.22 seconds to deallocate network for instance. [ 1201.436393] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.436767] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.436952] env[63175]: DEBUG nova.objects.instance [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'resources' on Instance uuid 92a9e153-06fb-420e-a96d-874ace510c5b {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.981535] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d9395a-1d5d-4bdf-9042-e030f52a981e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.988994] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c613f0-e5cd-4cdd-bb3e-36929f4a18ea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.018556] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635ddfa6-c22f-4b7d-9689-d21b99bff58c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.025510] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a95af3-c1c3-485d-8221-a5c99e164d53 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.038416] env[63175]: DEBUG nova.compute.provider_tree [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.541649] env[63175]: DEBUG nova.scheduler.client.report [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1203.046838] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.067601] env[63175]: INFO nova.scheduler.client.report [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted allocations for instance 92a9e153-06fb-420e-a96d-874ace510c5b [ 1203.576029] env[63175]: DEBUG oslo_concurrency.lockutils [None req-ad8b723d-a40b-4a98-bb52-fa343af14923 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "92a9e153-06fb-420e-a96d-874ace510c5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.477s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.316623] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.316893] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.822517] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1211.822517] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1211.822517] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1212.360874] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.361178] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.361229] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Forcefully refreshing network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1212.361538] env[63175]: DEBUG nova.objects.instance [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lazy-loading 'info_cache' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.090065] env[63175]: DEBUG nova.network.neutron [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.591933] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.592218] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updated the network info_cache for instance {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1214.592526] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.592756] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.592926] env[63175]: INFO nova.compute.manager [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Shelving [ 1214.594288] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.594661] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.594820] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.595024] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.595188] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.595340] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1214.595467] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1214.595605] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1215.099720] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.100103] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.100174] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.100269] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1215.101876] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edf79c5-4c55-4676-aa4c-c33c23e9b090 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.110823] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b372389-1198-406d-9e81-9d1a53eeead3 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.125776] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5c5a79-7985-438c-b39c-d94b3977efa5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.131787] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f91ea36-8674-45ee-ad85-c12224984172 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.160833] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181272MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1215.161055] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.161198] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.606172] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1215.606436] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50886c2e-e3d5-4f79-a8b9-0c3f476dfea0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.615651] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1215.615651] env[63175]: value = "task-1248723" [ 1215.615651] env[63175]: _type = "Task" [ 1215.615651] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.623380] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248723, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.124893] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248723, 'name': PowerOffVM_Task, 'duration_secs': 0.171622} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.125260] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1216.125919] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68dd648-c139-430f-8ff5-c26b8b31ee65 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.143335] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c50893-0f91-4e12-98ea-9f27b9f875a6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.184376] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Instance a6d70796-4b9c-4434-a721-dbd2f3ebce16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63175) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1216.184629] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1216.184795] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1216.210236] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0a2398-99d4-4f16-8270-d00c04862f00 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.217570] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ca4734-75b7-4a4b-b3f7-d02367aa3b51 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.248042] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc7f66d-56d8-40c7-9e70-253a1f66a5d8 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.254534] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db06e5f3-3ac9-4a9f-8e78-435bdb3c0b89 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.267180] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.652852] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Creating Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1216.653177] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6c745397-b701-4859-9be2-a9e2ffed4db9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.661532] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1216.661532] env[63175]: value = "task-1248724" [ 1216.661532] env[63175]: _type = "Task" [ 1216.661532] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.669515] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248724, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.769831] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1217.172204] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248724, 'name': CreateSnapshot_Task, 'duration_secs': 0.469687} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.172704] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Created Snapshot of the VM instance {{(pid=63175) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1217.173285] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cecc04-b46d-48a5-a132-402e59956c6f {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.275097] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1217.275375] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.114s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.691255] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Creating linked-clone VM from snapshot {{(pid=63175) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1217.691630] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6eb598b4-e1cd-4b5e-a411-8dd4bba6c574 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.700346] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1217.700346] env[63175]: value = "task-1248725" [ 1217.700346] env[63175]: _type = "Task" [ 1217.700346] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.708500] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248725, 'name': CloneVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.212874] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248725, 'name': CloneVM_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.710409] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248725, 'name': CloneVM_Task} progress is 100%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.212568] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248725, 'name': CloneVM_Task, 'duration_secs': 1.058345} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.212797] env[63175]: INFO nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Created linked-clone VM from snapshot [ 1219.213526] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01110d44-97ae-4922-a0e3-237fb4fa5a61 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.220346] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Uploading image 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1219.242562] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1219.242562] env[63175]: value = "vm-269123" [ 1219.242562] env[63175]: _type = "VirtualMachine" [ 1219.242562] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1219.242809] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bfd48d57-e564-4dbe-9836-3b87494a0e80 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.249276] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lease: (returnval){ [ 1219.249276] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52192302-edbf-1730-f504-ea4679125e6c" [ 1219.249276] env[63175]: _type = "HttpNfcLease" [ 1219.249276] env[63175]: } obtained for exporting VM: (result){ [ 1219.249276] env[63175]: value = "vm-269123" [ 1219.249276] env[63175]: _type = "VirtualMachine" [ 1219.249276] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1219.249537] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the lease: (returnval){ [ 1219.249537] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52192302-edbf-1730-f504-ea4679125e6c" [ 1219.249537] env[63175]: _type = "HttpNfcLease" [ 1219.249537] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1219.255621] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.255621] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52192302-edbf-1730-f504-ea4679125e6c" [ 1219.255621] env[63175]: _type = "HttpNfcLease" [ 1219.255621] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1219.757726] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.757726] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52192302-edbf-1730-f504-ea4679125e6c" [ 1219.757726] env[63175]: _type = "HttpNfcLease" [ 1219.757726] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1219.758031] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1219.758031] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52192302-edbf-1730-f504-ea4679125e6c" [ 1219.758031] env[63175]: _type = "HttpNfcLease" [ 1219.758031] env[63175]: }. {{(pid=63175) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1219.758716] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74fa604-2571-4270-b5e0-8d80f7007293 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.765684] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1219.765869] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk for reading. {{(pid=63175) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1219.852411] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3573b9de-4147-4742-aa6b-3a0408a75c82 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.198446] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1227.199404] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203873f5-8846-4575-863b-62a2f88c3529 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.205873] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1227.206057] env[63175]: ERROR oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk due to incomplete transfer. [ 1227.206275] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-78043acc-506b-4621-bb8c-79531dba095c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.213031] env[63175]: DEBUG oslo_vmware.rw_handles [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bc3272-6bc1-58fe-6db9-8ababa850b24/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1227.213235] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Uploaded image 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff to the Glance image server {{(pid=63175) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1227.215487] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Destroying the VM {{(pid=63175) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1227.215707] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c590e7e1-dc86-432c-a968-49c6a9bcca25 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.220779] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1227.220779] env[63175]: value = "task-1248727" [ 1227.220779] env[63175]: _type = "Task" [ 1227.220779] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.227902] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248727, 'name': Destroy_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.731151] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248727, 'name': Destroy_Task, 'duration_secs': 0.331053} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.731393] env[63175]: INFO nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Destroyed the VM [ 1227.731628] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleting Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1227.731874] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a9852e85-4792-4f5e-b9da-1356e678e068 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.737321] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1227.737321] env[63175]: value = "task-1248728" [ 1227.737321] env[63175]: _type = "Task" [ 1227.737321] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.744280] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248728, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.246326] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248728, 'name': RemoveSnapshot_Task, 'duration_secs': 0.342222} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.246747] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleted Snapshot of the VM instance {{(pid=63175) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1228.246847] env[63175]: DEBUG nova.compute.manager [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1228.247626] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0ddb66-72c5-4814-a78e-5fe1c115a863 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.758526] env[63175]: INFO nova.compute.manager [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Shelve offloading [ 1229.262381] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1229.262789] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0bc7ab3-9165-44d6-8d90-99ac84e31955 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.270925] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1229.270925] env[63175]: value = "task-1248729" [ 1229.270925] env[63175]: _type = "Task" [ 1229.270925] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.278622] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.781271] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] VM already powered off {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1229.781497] env[63175]: DEBUG nova.compute.manager [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1229.782262] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3703234-9577-4df6-bd4e-2b82f4f0a6ea {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.787457] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.787621] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.787787] env[63175]: DEBUG nova.network.neutron [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1230.468503] env[63175]: DEBUG nova.network.neutron [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.971618] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.180219] env[63175]: DEBUG nova.compute.manager [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-vif-unplugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1231.180451] env[63175]: DEBUG oslo_concurrency.lockutils [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.180663] env[63175]: DEBUG oslo_concurrency.lockutils [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.180838] env[63175]: DEBUG oslo_concurrency.lockutils [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.181013] env[63175]: DEBUG nova.compute.manager [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] No waiting events found dispatching network-vif-unplugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1231.181191] env[63175]: WARNING nova.compute.manager [req-471791ef-a2fb-42c0-90c7-33eacc0ccf47 req-ccbf17fc-c7d7-48b5-a39a-afa8be191a4b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received unexpected event network-vif-unplugged-21577a64-90af-48ef-b945-bb0bed6cc80e for instance with vm_state shelved and task_state shelving_offloading. [ 1231.203450] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1231.204307] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2fb1bb-f0e9-48eb-b047-be18b1044d69 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.211882] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1231.212122] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-493fabc4-d1d9-4894-8eda-2700ed76f7c1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.276643] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1231.276862] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1231.277057] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleting the datastore file [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1231.277353] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0571e3a5-b878-45ed-abf0-8f0d5e136f16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.283484] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1231.283484] env[63175]: value = "task-1248731" [ 1231.283484] env[63175]: _type = "Task" [ 1231.283484] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.290806] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.793139] env[63175]: DEBUG oslo_vmware.api [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130422} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.793431] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1231.793588] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1231.793768] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1231.817761] env[63175]: INFO nova.scheduler.client.report [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted allocations for instance a6d70796-4b9c-4434-a721-dbd2f3ebce16 [ 1232.322214] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.322214] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.322520] env[63175]: DEBUG nova.objects.instance [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'resources' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.825565] env[63175]: DEBUG nova.objects.instance [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'numa_topology' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.207689] env[63175]: DEBUG nova.compute.manager [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1233.207899] env[63175]: DEBUG nova.compute.manager [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing instance network info cache due to event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1233.208145] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.208316] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.208496] env[63175]: DEBUG nova.network.neutron [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1233.328062] env[63175]: DEBUG nova.objects.base [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63175) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1233.357443] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69259cec-5db4-4b2f-be7f-3ca6efbf9543 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.364846] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba6bbc2-086d-4e7b-844f-c16251b09b6d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.394444] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c125c38-14c8-4ebc-a865-4d74196b7415 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.400994] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab09263b-9001-4cbc-a267-d846e1fc7a16 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.414052] env[63175]: DEBUG nova.compute.provider_tree [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.481418] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.892171] env[63175]: DEBUG nova.network.neutron [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updated VIF entry in instance network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1233.892528] env[63175]: DEBUG nova.network.neutron [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": null, "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap21577a64-90", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1233.917381] env[63175]: DEBUG nova.scheduler.client.report [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1234.395777] env[63175]: DEBUG oslo_concurrency.lockutils [req-8f9fbff5-d95a-4901-bd17-e168ebc5b705 req-225a77ba-6052-479c-b918-f816b3175f64 service nova] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.421821] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.100s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.929663] env[63175]: DEBUG oslo_concurrency.lockutils [None req-8060621f-a3f6-4b3d-b331-f5251f4d0885 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.337s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.930561] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.449s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.930779] env[63175]: INFO nova.compute.manager [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Unshelving [ 1235.951045] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.951322] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.951536] env[63175]: DEBUG nova.objects.instance [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'pci_requests' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.455147] env[63175]: DEBUG nova.objects.instance [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'numa_topology' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.957613] env[63175]: INFO nova.compute.claims [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1237.989578] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f52d6ac-0642-4cfd-8db0-ebfb2e82499b {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.996853] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01d2476-f295-4ad5-80b0-0845bb282318 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.025068] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e3e4f3-a291-40bb-afa6-b149ce18f392 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.031603] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da23059f-b6bf-4e47-8499-04ea62a70cd7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.043887] env[63175]: DEBUG nova.compute.provider_tree [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1238.547282] env[63175]: DEBUG nova.scheduler.client.report [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1239.052782] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.101s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.079570] env[63175]: INFO nova.network.neutron [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating port 21577a64-90af-48ef-b945-bb0bed6cc80e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1240.432088] env[63175]: DEBUG nova.compute.manager [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1240.432336] env[63175]: DEBUG oslo_concurrency.lockutils [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.432562] env[63175]: DEBUG oslo_concurrency.lockutils [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.432678] env[63175]: DEBUG oslo_concurrency.lockutils [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.432848] env[63175]: DEBUG nova.compute.manager [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] No waiting events found dispatching network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1240.433022] env[63175]: WARNING nova.compute.manager [req-d66dc7a5-5ab3-4358-98c7-20eb196ce218 req-49afe0a8-21d3-4c59-80b5-403eebe4034b service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received unexpected event network-vif-plugged-21577a64-90af-48ef-b945-bb0bed6cc80e for instance with vm_state shelved_offloaded and task_state spawning. [ 1240.513834] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.515026] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.515026] env[63175]: DEBUG nova.network.neutron [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1241.191336] env[63175]: DEBUG nova.network.neutron [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.693842] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.720313] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T14:43:05Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a126cabb65dc8b4878244b1e59518b7b',container_format='bare',created_at=2024-10-10T14:56:18Z,direct_url=,disk_format='vmdk',id=7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-114097617-shelved',owner='ca6ed794d781428982398dcfb3998f08',properties=ImageMetaProps,protected=,size=31660544,status='active',tags=,updated_at=2024-10-10T14:56:32Z,virtual_size=,visibility=), allow threads: False {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1241.720631] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1241.720825] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image limits 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1241.721073] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Flavor pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1241.721271] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Image pref 0:0:0 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1241.721465] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63175) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1241.721710] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1241.721915] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1241.722147] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Got 1 possible topologies {{(pid=63175) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1241.722338] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1241.722553] env[63175]: DEBUG nova.virt.hardware [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63175) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1241.723482] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c137dd-fc56-4924-ac1c-1a20e3a26a09 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.731808] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba5b6cf-5e59-463e-a289-ff1436554c20 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.744710] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:06:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21577a64-90af-48ef-b945-bb0bed6cc80e', 'vif_model': 'vmxnet3'}] {{(pid=63175) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1241.751699] env[63175]: DEBUG oslo.service.loopingcall [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1241.751914] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Creating VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1241.752120] env[63175]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-114ef3e7-61b8-471d-ae95-3d56daf51787 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.769815] env[63175]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1241.769815] env[63175]: value = "task-1248732" [ 1241.769815] env[63175]: _type = "Task" [ 1241.769815] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.776648] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248732, 'name': CreateVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.280405] env[63175]: DEBUG oslo_vmware.api [-] Task: {'id': task-1248732, 'name': CreateVM_Task, 'duration_secs': 0.296308} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.280570] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Created VM on the ESX host {{(pid=63175) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1242.281245] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.281423] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.281856] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1242.282122] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13620b72-88a8-4a70-aaf4-1883929c19e6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.288020] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1242.288020] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5231d430-8b11-c970-af7a-255138299913" [ 1242.288020] env[63175]: _type = "Task" [ 1242.288020] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.293388] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]5231d430-8b11-c970-af7a-255138299913, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.456533] env[63175]: DEBUG nova.compute.manager [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1242.456743] env[63175]: DEBUG nova.compute.manager [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing instance network info cache due to event network-changed-21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1242.456971] env[63175]: DEBUG oslo_concurrency.lockutils [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.457132] env[63175]: DEBUG oslo_concurrency.lockutils [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.457314] env[63175]: DEBUG nova.network.neutron [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Refreshing network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1242.796164] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.796643] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Processing image 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1242.796901] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.797169] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.797463] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1242.797778] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26889f85-4fde-40da-b4ff-7ce76aac4a10 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.806681] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1242.806932] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63175) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1242.807686] env[63175]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c5ac32c-db89-4d82-9afe-e3d7567b3d64 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.812206] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1242.812206] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bdd7ec-5def-4e22-c0dc-04d781206d59" [ 1242.812206] env[63175]: _type = "Task" [ 1242.812206] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.820771] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52bdd7ec-5def-4e22-c0dc-04d781206d59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.316935] env[63175]: DEBUG nova.network.neutron [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updated VIF entry in instance network info cache for port 21577a64-90af-48ef-b945-bb0bed6cc80e. {{(pid=63175) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1243.317338] env[63175]: DEBUG nova.network.neutron [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.322283] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Preparing fetch location {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1243.322516] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Fetch image to [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4/OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4.vmdk {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1243.322700] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Downloading stream optimized image 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff to [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4/OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4.vmdk on the data store datastore1 as vApp {{(pid=63175) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1243.322872] env[63175]: DEBUG nova.virt.vmwareapi.images [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Downloading image file data 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff to the ESX as VM named 'OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4' {{(pid=63175) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1243.387151] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1243.387151] env[63175]: value = "resgroup-9" [ 1243.387151] env[63175]: _type = "ResourcePool" [ 1243.387151] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1243.387444] env[63175]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-77ba4cdc-1582-43d7-9bfc-6ca0069cf0dd {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.407232] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lease: (returnval){ [ 1243.407232] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52980ff8-cd67-418c-4389-46322473e324" [ 1243.407232] env[63175]: _type = "HttpNfcLease" [ 1243.407232] env[63175]: } obtained for vApp import into resource pool (val){ [ 1243.407232] env[63175]: value = "resgroup-9" [ 1243.407232] env[63175]: _type = "ResourcePool" [ 1243.407232] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1243.407503] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the lease: (returnval){ [ 1243.407503] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52980ff8-cd67-418c-4389-46322473e324" [ 1243.407503] env[63175]: _type = "HttpNfcLease" [ 1243.407503] env[63175]: } to be ready. {{(pid=63175) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1243.416259] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1243.416259] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52980ff8-cd67-418c-4389-46322473e324" [ 1243.416259] env[63175]: _type = "HttpNfcLease" [ 1243.416259] env[63175]: } is initializing. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1243.823353] env[63175]: DEBUG oslo_concurrency.lockutils [req-357a70d2-4a64-41d1-86e6-ac320bfa4dac req-10e71679-6463-4d76-9d38-f0e1817319fe service nova] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.915313] env[63175]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1243.915313] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52980ff8-cd67-418c-4389-46322473e324" [ 1243.915313] env[63175]: _type = "HttpNfcLease" [ 1243.915313] env[63175]: } is ready. {{(pid=63175) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1243.915603] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1243.915603] env[63175]: value = "session[522fb6ae-3352-22ee-bd4f-a9a11a41cb7e]52980ff8-cd67-418c-4389-46322473e324" [ 1243.915603] env[63175]: _type = "HttpNfcLease" [ 1243.915603] env[63175]: }. {{(pid=63175) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1243.916285] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd339f9-1b94-4511-b0cd-64e2f26b4c07 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.923270] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk from lease info. {{(pid=63175) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1243.923442] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating HTTP connection to write to file with size = 31660544 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk. {{(pid=63175) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1243.985662] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6f99d18a-8855-4e92-95cc-c65d886d3e13 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.065051] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Completed reading data from the image iterator. {{(pid=63175) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1245.065539] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1245.066251] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33938f82-17d3-40f9-bfaf-1fa23c43218d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.073163] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk is in state: ready. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1245.073377] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk. {{(pid=63175) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1245.073604] env[63175]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-2c1e5db9-f5cd-45ca-bcd6-3c8dfc0c1522 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.243950] env[63175]: DEBUG oslo_vmware.rw_handles [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52936ce8-cd84-d637-3501-299be97abaf4/disk-0.vmdk. {{(pid=63175) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1245.244198] env[63175]: INFO nova.virt.vmwareapi.images [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Downloaded image file data 7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff [ 1245.245101] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869593ac-88db-4ddb-96b0-974564401e37 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.260777] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7de0246c-e098-47c0-a004-52a1260677af {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.288654] env[63175]: INFO nova.virt.vmwareapi.images [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] The imported VM was unregistered [ 1245.291056] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Caching image {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1245.291300] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Creating directory with path [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1245.291565] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c22edaf-5fd7-4752-aba8-7058571f72a7 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.312934] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Created directory with path [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff {{(pid=63175) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1245.313127] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4/OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4.vmdk to [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk. {{(pid=63175) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1245.313368] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9e85acac-569a-4815-8550-7f01c626df3a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.319971] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1245.319971] env[63175]: value = "task-1248735" [ 1245.319971] env[63175]: _type = "Task" [ 1245.319971] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.327559] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.831884] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.332570] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.834275] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.334496] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.836059] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248735, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.207033} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.836059] env[63175]: INFO nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4/OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4.vmdk to [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk. [ 1247.836059] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Cleaning up location [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4 {{(pid=63175) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1247.836059] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_5ae61120-70c6-4010-88fb-70e6c924f8e4 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1247.836486] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b215ef99-895b-446a-8215-897b3f8e52c9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.842745] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1247.842745] env[63175]: value = "task-1248736" [ 1247.842745] env[63175]: _type = "Task" [ 1247.842745] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.850264] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.352625] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033928} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.352991] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1248.353096] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1248.353299] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk to [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1248.353548] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2de13310-6298-448b-84e0-9bd0fc6c975a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.359987] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1248.359987] env[63175]: value = "task-1248737" [ 1248.359987] env[63175]: _type = "Task" [ 1248.359987] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.366775] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.871256] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.371186] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.872761] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.373421] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.872075] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248737, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.095169} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.872290] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff/7c66c9e2-1d8a-4055-ba63-e8ea6a4cb1ff.vmdk to [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk {{(pid=63175) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1250.873041] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5515ee-6ff8-4f9f-a7e0-72d46434b68e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.893663] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1250.893881] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b25cd79a-00f3-4512-a8dc-5af86148cd85 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.912557] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1250.912557] env[63175]: value = "task-1248738" [ 1250.912557] env[63175]: _type = "Task" [ 1250.912557] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.919658] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248738, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.422239] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248738, 'name': ReconfigVM_Task, 'duration_secs': 0.258702} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.422577] env[63175]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Reconfigured VM instance instance-0000006b to attach disk [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16/a6d70796-4b9c-4434-a721-dbd2f3ebce16.vmdk or device None with type streamOptimized {{(pid=63175) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1251.423172] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1362acc8-4c47-435e-8ffa-c70826083e8a {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.428809] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1251.428809] env[63175]: value = "task-1248739" [ 1251.428809] env[63175]: _type = "Task" [ 1251.428809] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.436144] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248739, 'name': Rename_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.939837] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248739, 'name': Rename_Task, 'duration_secs': 0.188432} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.940133] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powering on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1251.940365] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-497b251b-afdd-48af-a10b-3ab0e36469da {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.947063] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1251.947063] env[63175]: value = "task-1248740" [ 1251.947063] env[63175]: _type = "Task" [ 1251.947063] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.954294] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.457067] env[63175]: DEBUG oslo_vmware.api [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248740, 'name': PowerOnVM_Task, 'duration_secs': 0.481508} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.457415] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powered on the VM {{(pid=63175) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1252.549324] env[63175]: DEBUG nova.compute.manager [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1252.550283] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5874f40a-01e7-41cb-9c6f-f27e3adacf8c {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.067442] env[63175]: DEBUG oslo_concurrency.lockutils [None req-1f87855f-9b55-4382-9af2-9ca926c4c081 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.137s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.796191] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd5b996-3f22-472f-94b4-40c0f2cb2f30 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.803165] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Suspending the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1254.803403] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-16de0a5e-5be9-4dd2-8c16-d7074f7d61a9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.809406] env[63175]: DEBUG oslo_vmware.api [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1254.809406] env[63175]: value = "task-1248741" [ 1254.809406] env[63175]: _type = "Task" [ 1254.809406] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.816771] env[63175]: DEBUG oslo_vmware.api [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248741, 'name': SuspendVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.319102] env[63175]: DEBUG oslo_vmware.api [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248741, 'name': SuspendVM_Task} progress is 79%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.819526] env[63175]: DEBUG oslo_vmware.api [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248741, 'name': SuspendVM_Task, 'duration_secs': 0.591329} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.819861] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Suspended the VM {{(pid=63175) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1255.820018] env[63175]: DEBUG nova.compute.manager [None req-0bf4c63d-5b6f-4a4c-9a79-42d846ee1a81 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1255.820771] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25f9e33-88ae-4eec-b5f0-94044fe6b311 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.168914] env[63175]: INFO nova.compute.manager [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Resuming [ 1257.169567] env[63175]: DEBUG nova.objects.instance [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'flavor' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.679749] env[63175]: DEBUG oslo_concurrency.lockutils [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1258.680214] env[63175]: DEBUG oslo_concurrency.lockutils [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquired lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.680214] env[63175]: DEBUG nova.network.neutron [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Building network info cache for instance {{(pid=63175) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1259.362628] env[63175]: DEBUG nova.network.neutron [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [{"id": "21577a64-90af-48ef-b945-bb0bed6cc80e", "address": "fa:16:3e:ec:06:e1", "network": {"id": "7ada5663-a441-4a77-8751-2d59ae49780b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-5134413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca6ed794d781428982398dcfb3998f08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21577a64-90", "ovs_interfaceid": "21577a64-90af-48ef-b945-bb0bed6cc80e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1259.865459] env[63175]: DEBUG oslo_concurrency.lockutils [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Releasing lock "refresh_cache-a6d70796-4b9c-4434-a721-dbd2f3ebce16" {{(pid=63175) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.866460] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530b0618-c609-423a-8a84-cf91178a89b9 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.873563] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Resuming the VM {{(pid=63175) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1259.873799] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08c111b7-2d18-495b-8f48-03d127719531 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.880608] env[63175]: DEBUG oslo_vmware.api [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1259.880608] env[63175]: value = "task-1248742" [ 1259.880608] env[63175]: _type = "Task" [ 1259.880608] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.887906] env[63175]: DEBUG oslo_vmware.api [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.392664] env[63175]: DEBUG oslo_vmware.api [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248742, 'name': PowerOnVM_Task, 'duration_secs': 0.50014} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.392973] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Resumed the VM {{(pid=63175) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1260.393199] env[63175]: DEBUG nova.compute.manager [None req-11455be1-28c4-465c-844d-1e74dab7b5a0 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Checking state {{(pid=63175) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1260.394021] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658c45f1-49e6-4e7a-b6a3-668bd2a0467d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.616565] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.616911] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.617061] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.617264] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.617464] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.619591] env[63175]: INFO nova.compute.manager [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Terminating instance [ 1266.124125] env[63175]: DEBUG nova.compute.manager [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Start destroying the instance on the hypervisor. {{(pid=63175) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1266.124357] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Destroying instance {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1266.125236] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d648405-119c-4724-ac63-c56736ac473e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.133232] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powering off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1266.133463] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d3cea12-0d19-4cbf-9752-db7fb016912e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.139785] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1266.139785] env[63175]: value = "task-1248743" [ 1266.139785] env[63175]: _type = "Task" [ 1266.139785] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.148595] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.650300] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248743, 'name': PowerOffVM_Task, 'duration_secs': 0.194682} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.650672] env[63175]: DEBUG nova.virt.vmwareapi.vm_util [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Powered off the VM {{(pid=63175) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1266.650672] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Unregistering the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1266.650905] env[63175]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36210631-c412-4325-a611-92989b175e3d {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.716322] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Unregistered the VM {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1266.716539] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleting contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1266.716727] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleting the datastore file [datastore1] a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1266.716993] env[63175]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51e6590e-112e-4f69-8a8d-9138179969fc {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.723468] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for the task: (returnval){ [ 1266.723468] env[63175]: value = "task-1248745" [ 1266.723468] env[63175]: _type = "Task" [ 1266.723468] env[63175]: } to complete. {{(pid=63175) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.731096] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.233724] env[63175]: DEBUG oslo_vmware.api [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Task: {'id': task-1248745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126749} completed successfully. {{(pid=63175) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.233980] env[63175]: DEBUG nova.virt.vmwareapi.ds_util [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted the datastore file {{(pid=63175) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1267.234188] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deleted contents of the VM from datastore datastore1 {{(pid=63175) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1267.234384] env[63175]: DEBUG nova.virt.vmwareapi.vmops [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Instance destroyed {{(pid=63175) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1267.234635] env[63175]: INFO nova.compute.manager [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1267.234913] env[63175]: DEBUG oslo.service.loopingcall [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63175) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1267.235130] env[63175]: DEBUG nova.compute.manager [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Deallocating network for instance {{(pid=63175) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1267.235227] env[63175]: DEBUG nova.network.neutron [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] deallocate_for_instance() {{(pid=63175) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1267.475684] env[63175]: DEBUG nova.compute.manager [req-65d19664-5c19-4b35-8c19-4bf55f0a7b26 req-213e5c0b-2ae2-498e-b6c3-3bcd76863f84 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Received event network-vif-deleted-21577a64-90af-48ef-b945-bb0bed6cc80e {{(pid=63175) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1267.475947] env[63175]: INFO nova.compute.manager [req-65d19664-5c19-4b35-8c19-4bf55f0a7b26 req-213e5c0b-2ae2-498e-b6c3-3bcd76863f84 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Neutron deleted interface 21577a64-90af-48ef-b945-bb0bed6cc80e; detaching it from the instance and deleting it from the info cache [ 1267.476192] env[63175]: DEBUG nova.network.neutron [req-65d19664-5c19-4b35-8c19-4bf55f0a7b26 req-213e5c0b-2ae2-498e-b6c3-3bcd76863f84 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1267.956456] env[63175]: DEBUG nova.network.neutron [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Updating instance_info_cache with network_info: [] {{(pid=63175) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1267.979137] env[63175]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c82e5990-bc2a-4471-bd7d-3b0b52de6490 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.989763] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d487b9c-0e5d-447e-b07e-9897d8918870 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.013039] env[63175]: DEBUG nova.compute.manager [req-65d19664-5c19-4b35-8c19-4bf55f0a7b26 req-213e5c0b-2ae2-498e-b6c3-3bcd76863f84 service nova] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Detach interface failed, port_id=21577a64-90af-48ef-b945-bb0bed6cc80e, reason: Instance a6d70796-4b9c-4434-a721-dbd2f3ebce16 could not be found. {{(pid=63175) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1268.459641] env[63175]: INFO nova.compute.manager [-] [instance: a6d70796-4b9c-4434-a721-dbd2f3ebce16] Took 1.22 seconds to deallocate network for instance. [ 1268.966348] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.966731] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.966891] env[63175]: DEBUG nova.objects.instance [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lazy-loading 'resources' on Instance uuid a6d70796-4b9c-4434-a721-dbd2f3ebce16 {{(pid=63175) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1269.501319] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929febf3-266c-44b3-bf00-4907a487670e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.509019] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45480b8d-9d3d-4f6c-838a-d9e1bef38801 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.539039] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e793a4f8-bd85-4b3c-a89a-866808a24457 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.546016] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5521792-ae56-4b18-b681-206c8ad0ff71 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.558895] env[63175]: DEBUG nova.compute.provider_tree [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.062359] env[63175]: DEBUG nova.scheduler.client.report [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1270.567914] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.585307] env[63175]: INFO nova.scheduler.client.report [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Deleted allocations for instance a6d70796-4b9c-4434-a721-dbd2f3ebce16 [ 1271.093589] env[63175]: DEBUG oslo_concurrency.lockutils [None req-257e9151-b2db-4373-80a0-7490f3686084 tempest-ServersNegativeTestJSON-1945716721 tempest-ServersNegativeTestJSON-1945716721-project-member] Lock "a6d70796-4b9c-4434-a721-dbd2f3ebce16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.477s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.277090] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.277499] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.277499] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Starting heal instance info cache {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1277.277649] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Rebuilding the list of instances to heal {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1277.780906] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Didn't find any instances for network info cache update. {{(pid=63175) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1277.781247] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.781374] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.781527] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.781677] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.781825] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.781966] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1277.782116] env[63175]: DEBUG nova.compute.manager [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63175) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1277.782286] env[63175]: DEBUG oslo_service.periodic_task [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Running periodic task ComputeManager.update_available_resource {{(pid=63175) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.285901] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.286281] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.286323] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.286469] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63175) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1278.287383] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24163355-1416-4d07-872c-25fc083cf8d0 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.295698] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51dde40-b6f3-4264-a26b-09581b744ca5 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.310378] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212a9c35-ef9e-467c-84d7-951f4ab2a0e1 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.316917] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995fd9a5-f27e-45e8-975f-5ab955bb640e {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.345080] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181319MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=63175) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1278.345265] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.345430] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.365038] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1279.365282] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63175) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1279.378414] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e4615f-895a-42f4-b4dc-238c650db867 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.386011] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f091d7d-8c03-489f-96df-685f02179cf6 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.415173] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106fb5fc-179b-4658-bdc4-8291529455ee {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.421689] env[63175]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95bafae-6a4c-4b8d-8c14-3ca775505bc4 {{(pid=63175) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.074553] env[63175]: DEBUG nova.compute.provider_tree [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed in ProviderTree for provider: f10bd0b1-85ed-4db3-a7f2-6e16e785219e {{(pid=63175) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.578114] env[63175]: DEBUG nova.scheduler.client.report [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Inventory has not changed for provider f10bd0b1-85ed-4db3-a7f2-6e16e785219e based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63175) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1281.083294] env[63175]: DEBUG nova.compute.resource_tracker [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63175) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1281.083546] env[63175]: DEBUG oslo_concurrency.lockutils [None req-91aae357-ca3e-4be6-8153-5016513ff43d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.738s {{(pid=63175) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}